var/home/core/zuul-output/0000755000175000017500000000000015067747417014547 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067774616015514 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006506433615067774606017734 0ustar rootrootOct 03 13:30:29 crc systemd[1]: Starting Kubernetes Kubelet... Oct 03 13:30:29 crc restorecon[4738]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:29 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 13:30:30 crc restorecon[4738]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 03 13:30:31 crc kubenswrapper[4959]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.382441 4959 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393868 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393908 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393919 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393929 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393938 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393947 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393956 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393963 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393971 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393979 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393987 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.393994 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394003 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394011 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394019 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394027 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394034 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394042 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394061 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394069 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394077 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394085 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394093 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394101 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394109 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394116 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394124 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394131 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394139 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394150 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394161 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394171 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394180 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394225 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394237 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394248 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394257 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394265 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394275 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394284 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394292 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394301 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394308 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394316 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394324 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394333 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394342 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394349 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394357 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394366 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394373 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394381 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394389 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394396 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394404 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394411 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394419 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394427 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394434 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394442 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394450 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394458 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394465 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394473 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394480 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394490 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394498 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394507 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394518 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394528 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.394537 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394695 4959 flags.go:64] FLAG: --address="0.0.0.0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394718 4959 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394735 4959 flags.go:64] FLAG: --anonymous-auth="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394748 4959 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394761 4959 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394771 4959 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394783 4959 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394795 4959 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394804 4959 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394813 4959 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394823 4959 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394834 4959 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394843 4959 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394852 4959 flags.go:64] FLAG: --cgroup-root="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394861 4959 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394870 4959 flags.go:64] FLAG: --client-ca-file="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394879 4959 flags.go:64] FLAG: --cloud-config="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394888 4959 flags.go:64] FLAG: --cloud-provider="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394897 4959 flags.go:64] FLAG: --cluster-dns="[]" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394909 4959 flags.go:64] FLAG: --cluster-domain="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394918 4959 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394927 4959 flags.go:64] FLAG: --config-dir="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394936 4959 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394946 4959 flags.go:64] FLAG: --container-log-max-files="5" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394956 4959 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394965 4959 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394975 4959 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394985 4959 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.394994 4959 flags.go:64] FLAG: --contention-profiling="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395003 4959 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395012 4959 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395022 4959 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395033 4959 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395044 4959 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395053 4959 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395062 4959 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395070 4959 flags.go:64] FLAG: --enable-load-reader="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395079 4959 flags.go:64] FLAG: --enable-server="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395088 4959 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395099 4959 flags.go:64] FLAG: --event-burst="100" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395108 4959 flags.go:64] FLAG: --event-qps="50" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395117 4959 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395128 4959 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395137 4959 flags.go:64] FLAG: --eviction-hard="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395147 4959 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395156 4959 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395165 4959 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395174 4959 flags.go:64] FLAG: --eviction-soft="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395184 4959 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395231 4959 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395244 4959 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395255 4959 flags.go:64] FLAG: --experimental-mounter-path="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395266 4959 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395278 4959 flags.go:64] FLAG: --fail-swap-on="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395288 4959 flags.go:64] FLAG: --feature-gates="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395299 4959 flags.go:64] FLAG: --file-check-frequency="20s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395310 4959 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395319 4959 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395329 4959 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395338 4959 flags.go:64] FLAG: --healthz-port="10248" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395348 4959 flags.go:64] FLAG: --help="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395357 4959 flags.go:64] FLAG: --hostname-override="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395365 4959 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395375 4959 flags.go:64] FLAG: --http-check-frequency="20s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395384 4959 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395393 4959 flags.go:64] FLAG: --image-credential-provider-config="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395402 4959 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395411 4959 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395432 4959 flags.go:64] FLAG: --image-service-endpoint="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395441 4959 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395451 4959 flags.go:64] FLAG: --kube-api-burst="100" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395460 4959 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395469 4959 flags.go:64] FLAG: --kube-api-qps="50" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395480 4959 flags.go:64] FLAG: --kube-reserved="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395489 4959 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395498 4959 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395507 4959 flags.go:64] FLAG: --kubelet-cgroups="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395516 4959 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395525 4959 flags.go:64] FLAG: --lock-file="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395534 4959 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395543 4959 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395552 4959 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395566 4959 flags.go:64] FLAG: --log-json-split-stream="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395575 4959 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395585 4959 flags.go:64] FLAG: --log-text-split-stream="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395594 4959 flags.go:64] FLAG: --logging-format="text" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395603 4959 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395614 4959 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395623 4959 flags.go:64] FLAG: --manifest-url="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395632 4959 flags.go:64] FLAG: --manifest-url-header="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395644 4959 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395653 4959 flags.go:64] FLAG: --max-open-files="1000000" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395664 4959 flags.go:64] FLAG: --max-pods="110" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395673 4959 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395682 4959 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395692 4959 flags.go:64] FLAG: --memory-manager-policy="None" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395701 4959 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395710 4959 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395719 4959 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395728 4959 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395747 4959 flags.go:64] FLAG: --node-status-max-images="50" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395756 4959 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395765 4959 flags.go:64] FLAG: --oom-score-adj="-999" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395775 4959 flags.go:64] FLAG: --pod-cidr="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395785 4959 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395799 4959 flags.go:64] FLAG: --pod-manifest-path="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395808 4959 flags.go:64] FLAG: --pod-max-pids="-1" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395817 4959 flags.go:64] FLAG: --pods-per-core="0" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395826 4959 flags.go:64] FLAG: --port="10250" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395835 4959 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395844 4959 flags.go:64] FLAG: --provider-id="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395853 4959 flags.go:64] FLAG: --qos-reserved="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395861 4959 flags.go:64] FLAG: --read-only-port="10255" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395870 4959 flags.go:64] FLAG: --register-node="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395879 4959 flags.go:64] FLAG: --register-schedulable="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395889 4959 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395903 4959 flags.go:64] FLAG: --registry-burst="10" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395912 4959 flags.go:64] FLAG: --registry-qps="5" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395921 4959 flags.go:64] FLAG: --reserved-cpus="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395930 4959 flags.go:64] FLAG: --reserved-memory="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395941 4959 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395950 4959 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395960 4959 flags.go:64] FLAG: --rotate-certificates="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395968 4959 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395977 4959 flags.go:64] FLAG: --runonce="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395987 4959 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.395996 4959 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396005 4959 flags.go:64] FLAG: --seccomp-default="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396014 4959 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396023 4959 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396033 4959 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396043 4959 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396052 4959 flags.go:64] FLAG: --storage-driver-password="root" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396061 4959 flags.go:64] FLAG: --storage-driver-secure="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396070 4959 flags.go:64] FLAG: --storage-driver-table="stats" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396079 4959 flags.go:64] FLAG: --storage-driver-user="root" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396088 4959 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396098 4959 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396107 4959 flags.go:64] FLAG: --system-cgroups="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396116 4959 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396132 4959 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396141 4959 flags.go:64] FLAG: --tls-cert-file="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396150 4959 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396162 4959 flags.go:64] FLAG: --tls-min-version="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396171 4959 flags.go:64] FLAG: --tls-private-key-file="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396179 4959 flags.go:64] FLAG: --topology-manager-policy="none" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396188 4959 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396227 4959 flags.go:64] FLAG: --topology-manager-scope="container" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396239 4959 flags.go:64] FLAG: --v="2" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396285 4959 flags.go:64] FLAG: --version="false" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396300 4959 flags.go:64] FLAG: --vmodule="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396312 4959 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.396322 4959 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396575 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396589 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396597 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396605 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396615 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396622 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396630 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396645 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396652 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396664 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396677 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396690 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396701 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396714 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396727 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396737 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396747 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396757 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396767 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396777 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396786 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396796 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396806 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396819 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396828 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396839 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396849 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396858 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396868 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396877 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396887 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396898 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396907 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396918 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396927 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396941 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396953 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396965 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396976 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.396991 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397002 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397012 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397023 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397033 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397043 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397053 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397063 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397072 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397081 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397090 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397100 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397110 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397119 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397128 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397137 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397147 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397157 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397166 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397176 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397232 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397245 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397255 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397265 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397275 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397284 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397294 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397304 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397313 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397326 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397338 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.397349 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.397382 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.408141 4959 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.408222 4959 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408371 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408385 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408394 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408405 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408414 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408422 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408433 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408445 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408455 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408464 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408473 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408481 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408488 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408496 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408504 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408552 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408560 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408569 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408577 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408587 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408599 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408608 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408619 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408629 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408638 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408647 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408656 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408666 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408675 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408686 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408695 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408704 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408712 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408720 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408730 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408738 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408746 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408754 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408764 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408773 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408782 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408790 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408800 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408809 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408817 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408826 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408838 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408848 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408885 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408895 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408905 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408915 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408925 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408935 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408944 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408956 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408965 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408974 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408984 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.408994 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409004 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409013 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409025 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409034 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409044 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409054 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409063 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409073 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409082 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409092 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409104 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.409122 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409477 4959 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409503 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409517 4959 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409533 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409543 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409553 4959 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409561 4959 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409569 4959 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409577 4959 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409586 4959 feature_gate.go:330] unrecognized feature gate: Example Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409595 4959 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409604 4959 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409613 4959 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409621 4959 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409629 4959 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409637 4959 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409644 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409652 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409661 4959 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409669 4959 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409677 4959 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409684 4959 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409692 4959 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409700 4959 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409708 4959 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409715 4959 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409724 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409733 4959 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409743 4959 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409753 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409765 4959 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409774 4959 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409784 4959 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409792 4959 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409801 4959 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409810 4959 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409817 4959 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409825 4959 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409835 4959 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409844 4959 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409853 4959 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409863 4959 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409873 4959 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409882 4959 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409890 4959 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409899 4959 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409907 4959 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409916 4959 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409924 4959 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409932 4959 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409942 4959 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409949 4959 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409957 4959 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409965 4959 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409973 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409981 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409989 4959 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.409996 4959 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410004 4959 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410014 4959 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410023 4959 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410032 4959 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410040 4959 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410048 4959 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410056 4959 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410063 4959 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410072 4959 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410079 4959 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410087 4959 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410095 4959 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.410104 4959 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.410118 4959 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.411730 4959 server.go:940] "Client rotation is on, will bootstrap in background" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.423546 4959 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.423666 4959 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.425898 4959 server.go:997] "Starting client certificate rotation" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.425941 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.426141 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-24 08:54:05.700166184 +0000 UTC Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.426276 4959 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1243h23m34.27389338s for next certificate rotation Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.492536 4959 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.505116 4959 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.528436 4959 log.go:25] "Validated CRI v1 runtime API" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.576621 4959 log.go:25] "Validated CRI v1 image API" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.578474 4959 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.587095 4959 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-03-13-25-11-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.587336 4959 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.607555 4959 manager.go:217] Machine: {Timestamp:2025-10-03 13:30:31.602558596 +0000 UTC m=+0.805902033 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:168d26af-4a20-4818-904a-4d599a720a2a BootID:e9ec5ca6-9088-4c22-bd9a-b3c75487e558 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:29:cc:5a Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:29:cc:5a Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:14:1f:30 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1e:04:44 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:47:3e:61 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d9:72:6d Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:c1:83:2d Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:7a:87:9e:ff:37 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:c6:74:9b:92:23:be Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.608006 4959 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.608281 4959 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.610733 4959 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.611058 4959 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.611113 4959 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.611468 4959 topology_manager.go:138] "Creating topology manager with none policy" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.611488 4959 container_manager_linux.go:303] "Creating device plugin manager" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.612090 4959 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.612144 4959 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.612438 4959 state_mem.go:36] "Initialized new in-memory state store" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.612572 4959 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.618033 4959 kubelet.go:418] "Attempting to sync node with API server" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.618069 4959 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.618098 4959 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.618120 4959 kubelet.go:324] "Adding apiserver pod source" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.618139 4959 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.622515 4959 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.623706 4959 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.625480 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.625578 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.625661 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.625688 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.626101 4959 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627566 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627592 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627602 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627610 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627624 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627632 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627640 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627653 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627662 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627670 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627699 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.627708 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.629626 4959 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.630527 4959 server.go:1280] "Started kubelet" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.630835 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:31 crc systemd[1]: Started Kubernetes Kubelet. Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.632338 4959 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.632467 4959 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.632791 4959 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633457 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633490 4959 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633720 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 17:17:33.922071209 +0000 UTC Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633756 4959 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1107h47m2.288317352s for next certificate rotation Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633768 4959 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633755 4959 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.633805 4959 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.633802 4959 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.634234 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="200ms" Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.634655 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.634719 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.635206 4959 factory.go:55] Registering systemd factory Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.635231 4959 factory.go:221] Registration of the systemd container factory successfully Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.636615 4959 server.go:460] "Adding debug handlers to kubelet server" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.638469 4959 factory.go:153] Registering CRI-O factory Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.638507 4959 factory.go:221] Registration of the crio container factory successfully Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.640004 4959 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.640063 4959 factory.go:103] Registering Raw factory Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.640299 4959 manager.go:1196] Started watching for new ooms in manager Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.636756 4959 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.245:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186afe51c73d8dfb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 13:30:31.630106107 +0000 UTC m=+0.833449544,LastTimestamp:2025-10-03 13:30:31.630106107 +0000 UTC m=+0.833449544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.642585 4959 manager.go:319] Starting recovery of all containers Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647437 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647520 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647538 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647553 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647567 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647579 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647592 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647608 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647625 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647639 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647652 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647666 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647681 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647702 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647717 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647733 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647749 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647764 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647803 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647815 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647827 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647841 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647853 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647864 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647875 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647885 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647896 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647912 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647922 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647934 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647943 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647953 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647964 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647975 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.647994 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648037 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648050 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648064 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648078 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648089 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648104 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648118 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648174 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648185 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648213 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648227 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648239 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648250 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648261 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648273 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648285 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648296 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648335 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648347 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648361 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648373 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648385 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648396 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648407 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648418 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648429 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648442 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648455 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648469 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648483 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648498 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648513 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648528 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648545 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648607 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648620 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648634 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648646 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648659 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648673 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648688 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648703 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648720 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648736 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648750 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648765 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648779 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648793 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648808 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648822 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648834 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648847 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648863 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648881 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648895 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648910 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648922 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648934 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648943 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648954 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648964 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648973 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648983 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.648994 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649004 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649013 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649023 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649032 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649042 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649058 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649069 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649079 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649090 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649101 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649111 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649121 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649131 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649141 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649151 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649161 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649172 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649181 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649210 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649219 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649229 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649239 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649250 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649260 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649269 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649278 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649287 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649296 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649306 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649321 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649650 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649667 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649679 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649688 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649699 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649709 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649718 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649728 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649737 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649746 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649756 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649765 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649778 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649789 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649799 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649812 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649822 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649831 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649841 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649853 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649864 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649874 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649886 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649897 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649908 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649919 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649930 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649940 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649953 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649965 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649978 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.649995 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650006 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650016 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650026 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650043 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650053 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650063 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650073 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650083 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650093 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650104 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650114 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.650125 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652208 4959 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652238 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652251 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652266 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652276 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652287 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652302 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652312 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652322 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652334 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652344 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652357 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652371 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652385 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652397 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652412 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652426 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652441 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652456 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652471 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652485 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652497 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652510 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652522 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652537 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652555 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652571 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652587 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652602 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652617 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652631 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652645 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652660 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652675 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652689 4959 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652700 4959 reconstruct.go:97] "Volume reconstruction finished" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.652708 4959 reconciler.go:26] "Reconciler: start to sync state" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.661147 4959 manager.go:324] Recovery completed Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.671068 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.672560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.672603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.672611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.673307 4959 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.673331 4959 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.673351 4959 state_mem.go:36] "Initialized new in-memory state store" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.681512 4959 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.684120 4959 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.684179 4959 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.684225 4959 kubelet.go:2335] "Starting kubelet main sync loop" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.684279 4959 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 03 13:30:31 crc kubenswrapper[4959]: W1003 13:30:31.685012 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.685065 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.690930 4959 policy_none.go:49] "None policy: Start" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.692311 4959 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.692353 4959 state_mem.go:35] "Initializing new in-memory state store" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.734221 4959 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.766799 4959 manager.go:334] "Starting Device Plugin manager" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.767107 4959 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.767129 4959 server.go:79] "Starting device plugin registration server" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.767993 4959 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.768017 4959 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.768246 4959 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.768367 4959 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.768377 4959 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.777245 4959 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.784562 4959 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.784680 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788594 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788887 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.788951 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789598 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789735 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.789808 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790561 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790677 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790937 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.790979 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792803 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792829 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792912 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.792933 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793750 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.793768 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.794318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.794348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.794361 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.835242 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="400ms" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854591 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854612 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854630 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854656 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854675 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854862 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.854962 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.855017 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.855053 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.855083 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.869058 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.870379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.870422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.870435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.870464 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:31 crc kubenswrapper[4959]: E1003 13:30:31.870924 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956625 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956724 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956774 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956815 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956887 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956914 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.956916 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957043 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957039 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957102 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957116 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957139 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957108 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957061 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957416 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957503 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957525 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957557 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957614 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957663 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957672 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957748 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957736 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957779 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957821 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 13:30:31 crc kubenswrapper[4959]: I1003 13:30:31.957870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.072116 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.073481 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.073541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.073559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.073588 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.074104 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.125132 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.133080 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.155495 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.177992 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.183635 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.207978 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-afb9a07899f0dbaf50d924831bda4355056700496b6a100c414d788079dc293a WatchSource:0}: Error finding container afb9a07899f0dbaf50d924831bda4355056700496b6a100c414d788079dc293a: Status 404 returned error can't find the container with id afb9a07899f0dbaf50d924831bda4355056700496b6a100c414d788079dc293a Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.220440 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-e3226befed9269305ffc038b2f31013f1a25a05a05dcedf90136e4068a7eca21 WatchSource:0}: Error finding container e3226befed9269305ffc038b2f31013f1a25a05a05dcedf90136e4068a7eca21: Status 404 returned error can't find the container with id e3226befed9269305ffc038b2f31013f1a25a05a05dcedf90136e4068a7eca21 Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.227094 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2345a0905680d9be68445ce0f11c67ce97523856b46f65d8bce4d40d855e4680 WatchSource:0}: Error finding container 2345a0905680d9be68445ce0f11c67ce97523856b46f65d8bce4d40d855e4680: Status 404 returned error can't find the container with id 2345a0905680d9be68445ce0f11c67ce97523856b46f65d8bce4d40d855e4680 Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.228728 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-db2590d160d3d5ed5cb50382c1027601b15ded67ef734f89824ed56253466ec9 WatchSource:0}: Error finding container db2590d160d3d5ed5cb50382c1027601b15ded67ef734f89824ed56253466ec9: Status 404 returned error can't find the container with id db2590d160d3d5ed5cb50382c1027601b15ded67ef734f89824ed56253466ec9 Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.236457 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="800ms" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.475172 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.476767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.476822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.476834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.476867 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.477440 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.527297 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.527388 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.631972 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.689046 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"afb9a07899f0dbaf50d924831bda4355056700496b6a100c414d788079dc293a"} Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.690315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"db2590d160d3d5ed5cb50382c1027601b15ded67ef734f89824ed56253466ec9"} Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.691423 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2345a0905680d9be68445ce0f11c67ce97523856b46f65d8bce4d40d855e4680"} Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.692337 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e3226befed9269305ffc038b2f31013f1a25a05a05dcedf90136e4068a7eca21"} Oct 03 13:30:32 crc kubenswrapper[4959]: I1003 13:30:32.693085 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d907e76bc49e3870150e559b21682ddbbb2ea600b61f46df1fc13be8264dbf3d"} Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.855230 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.855320 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:32 crc kubenswrapper[4959]: W1003 13:30:32.922944 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:32 crc kubenswrapper[4959]: E1003 13:30:32.923038 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:33 crc kubenswrapper[4959]: W1003 13:30:33.027495 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:33 crc kubenswrapper[4959]: E1003 13:30:33.027594 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:33 crc kubenswrapper[4959]: E1003 13:30:33.037743 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="1.6s" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.278245 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.279630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.279674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.279686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.279713 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:33 crc kubenswrapper[4959]: E1003 13:30:33.280289 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.632374 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.696449 4959 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8" exitCode=0 Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.696516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.696671 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698375 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698546 4959 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765" exitCode=0 Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698639 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.698634 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.699447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.699466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.699476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.700430 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.700470 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.702468 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840" exitCode=0 Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.702554 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.702660 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.703549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.703587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.703599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.705011 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da" exitCode=0 Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.705044 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da"} Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.705175 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.706652 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.706695 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.706700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.706714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.707538 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.707585 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:33 crc kubenswrapper[4959]: I1003 13:30:33.707593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: W1003 13:30:34.138852 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:34 crc kubenswrapper[4959]: E1003 13:30:34.139002 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:34 crc kubenswrapper[4959]: W1003 13:30:34.623472 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:34 crc kubenswrapper[4959]: E1003 13:30:34.623541 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.633028 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:34 crc kubenswrapper[4959]: E1003 13:30:34.639352 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.245:6443: connect: connection refused" interval="3.2s" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.709748 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.709804 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.709816 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.709836 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.710716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.710750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.710760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.711778 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.711809 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.711902 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.712924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.712953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.712964 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714761 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714799 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714819 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714841 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.714936 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.715801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.715862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.715874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.717395 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd" exitCode=0 Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.717455 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.717550 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.718221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.718247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.718258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.736788 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8"} Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.736848 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.737734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.737765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.737774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.827105 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.880640 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.881881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.881925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.881937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:34 crc kubenswrapper[4959]: I1003 13:30:34.881965 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:34 crc kubenswrapper[4959]: E1003 13:30:34.882488 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.245:6443: connect: connection refused" node="crc" Oct 03 13:30:34 crc kubenswrapper[4959]: W1003 13:30:34.897308 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.245:6443: connect: connection refused Oct 03 13:30:34 crc kubenswrapper[4959]: E1003 13:30:34.897405 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.245:6443: connect: connection refused" logger="UnhandledError" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743083 4959 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8" exitCode=0 Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743279 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743318 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743355 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8"} Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743382 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743536 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743616 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.743427 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.744909 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.745173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.746119 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.746187 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.746246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:35 crc kubenswrapper[4959]: I1003 13:30:35.826107 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a"} Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750669 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f"} Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416"} Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750700 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1"} Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750711 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965"} Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750732 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750785 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.750732 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.751930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.751974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.751989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.751972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.752090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.752109 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.752858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.752884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:36 crc kubenswrapper[4959]: I1003 13:30:36.752893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.753686 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.755072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.755111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.755122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.768452 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.768609 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.769801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.769828 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:37 crc kubenswrapper[4959]: I1003 13:30:37.769838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.083262 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.084866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.084920 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.084932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.084964 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.396548 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.681591 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.688272 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.756684 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.756707 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758723 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.758803 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.818354 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.818547 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.818588 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.819539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.819588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:38 crc kubenswrapper[4959]: I1003 13:30:38.819602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.278166 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.479460 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.759760 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.759760 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:39 crc kubenswrapper[4959]: I1003 13:30:39.761657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:40 crc kubenswrapper[4959]: I1003 13:30:40.763057 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:40 crc kubenswrapper[4959]: I1003 13:30:40.764110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:40 crc kubenswrapper[4959]: I1003 13:30:40.764157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:40 crc kubenswrapper[4959]: I1003 13:30:40.764166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:41 crc kubenswrapper[4959]: E1003 13:30:41.777387 4959 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.100210 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.100403 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.101927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.101983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.102000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.274813 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.275054 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.276834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.276889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.276906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.280631 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.772413 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.773431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.773468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:44 crc kubenswrapper[4959]: I1003 13:30:44.773484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:45 crc kubenswrapper[4959]: E1003 13:30:45.316307 4959 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.186afe51c73d8dfb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 13:30:31.630106107 +0000 UTC m=+0.833449544,LastTimestamp:2025-10-03 13:30:31.630106107 +0000 UTC m=+0.833449544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.632544 4959 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 03 13:30:45 crc kubenswrapper[4959]: W1003 13:30:45.664098 4959 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.664181 4959 trace.go:236] Trace[146850363]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 13:30:35.662) (total time: 10001ms): Oct 03 13:30:45 crc kubenswrapper[4959]: Trace[146850363]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:30:45.664) Oct 03 13:30:45 crc kubenswrapper[4959]: Trace[146850363]: [10.001684321s] [10.001684321s] END Oct 03 13:30:45 crc kubenswrapper[4959]: E1003 13:30:45.664221 4959 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.725920 4959 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.726002 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.735222 4959 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.735290 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.776068 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.780373 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844" exitCode=255 Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.780424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844"} Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.780587 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.781364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.781397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.781405 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:45 crc kubenswrapper[4959]: I1003 13:30:45.781941 4959 scope.go:117] "RemoveContainer" containerID="2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844" Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.784429 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.787001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e"} Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.787267 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.788554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.788610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:46 crc kubenswrapper[4959]: I1003 13:30:46.788624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:47 crc kubenswrapper[4959]: I1003 13:30:47.274945 4959 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 13:30:47 crc kubenswrapper[4959]: I1003 13:30:47.275056 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.823481 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.823655 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.823771 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.824872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.824905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.824918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:48 crc kubenswrapper[4959]: I1003 13:30:48.827767 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.071401 4959 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.626178 4959 apiserver.go:52] "Watching apiserver" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.631658 4959 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.632130 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.632759 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.632860 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.632953 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.632989 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:49 crc kubenswrapper[4959]: E1003 13:30:49.633104 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.633148 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.633209 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:49 crc kubenswrapper[4959]: E1003 13:30:49.633345 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:49 crc kubenswrapper[4959]: E1003 13:30:49.633339 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.634376 4959 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.635923 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.636418 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.636660 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.636885 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.637168 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.637311 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.637387 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.637481 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.637563 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.660397 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.678079 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.693671 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.709670 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.720938 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.736306 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.751480 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.761980 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:49 crc kubenswrapper[4959]: I1003 13:30:49.810694 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.685333 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.685505 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.727518 4959 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.728856 4959 trace.go:236] Trace[623829095]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 13:30:38.783) (total time: 11945ms): Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[623829095]: ---"Objects listed" error: 11945ms (13:30:50.728) Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[623829095]: [11.945787992s] [11.945787992s] END Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.728878 4959 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.729150 4959 trace.go:236] Trace[1215284224]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 13:30:39.452) (total time: 11276ms): Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[1215284224]: ---"Objects listed" error: 11275ms (13:30:50.728) Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[1215284224]: [11.276124855s] [11.276124855s] END Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.729392 4959 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.730498 4959 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.731022 4959 trace.go:236] Trace[184259682]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 13:30:39.321) (total time: 11409ms): Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[184259682]: ---"Objects listed" error: 11409ms (13:30:50.730) Oct 03 13:30:50 crc kubenswrapper[4959]: Trace[184259682]: [11.409948244s] [11.409948244s] END Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.731044 4959 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.731359 4959 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.831861 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832160 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832310 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832383 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832600 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832701 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832808 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832893 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.832965 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833070 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833160 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833304 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833408 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833081 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833116 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833526 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833568 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833504 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833701 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833209 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833214 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833479 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833492 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833621 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833698 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833734 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833873 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833906 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833931 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833932 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.833984 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834011 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834037 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834061 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834090 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834114 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834162 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834210 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834241 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834270 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834297 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834137 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834323 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834245 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834352 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834381 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834407 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834431 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834458 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834486 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834589 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834615 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834685 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834714 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834741 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834763 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834813 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834835 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834863 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834926 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835050 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835079 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835106 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835132 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835188 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835233 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835257 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835279 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835371 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835397 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835777 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835850 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835878 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835906 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835930 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836044 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836111 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836182 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836251 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836271 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836296 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836318 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836365 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836387 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836463 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836485 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836508 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836530 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836554 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836575 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836631 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836714 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836740 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836763 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836782 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836824 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836845 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836867 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836888 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836910 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836932 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836973 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836996 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837016 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837038 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837060 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837106 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837126 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837149 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837171 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837313 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837338 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837360 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837380 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837399 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837508 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837530 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837553 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837576 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837619 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837641 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837688 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837709 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837733 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837753 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837800 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837823 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837844 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837865 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837907 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837924 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837940 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837957 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837974 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837991 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838007 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838023 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838039 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838060 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838077 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838094 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838113 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838130 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838167 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838187 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838243 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838297 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838326 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838356 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838378 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838401 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838418 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838434 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838451 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838468 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838485 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838503 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838526 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838548 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838574 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838593 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838610 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838626 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838642 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838659 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838679 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838695 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838713 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838732 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838750 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838767 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838783 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838801 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838818 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838835 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838851 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838869 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838887 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838905 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838940 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838964 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838986 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839068 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839087 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839104 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839143 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839160 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839224 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839243 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839317 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839328 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839339 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839350 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839367 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839385 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839402 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839418 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839432 4959 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839446 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839460 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839473 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839483 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839493 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839502 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839511 4959 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841104 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834281 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834296 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834395 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834389 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834470 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834632 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834699 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834709 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834933 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834947 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.834959 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835054 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835212 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835252 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835258 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835431 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835483 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835641 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835652 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835757 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.835858 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836016 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836096 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836226 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836393 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.836494 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837050 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837361 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837460 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837620 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837689 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837781 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837805 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837884 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.837972 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838023 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838055 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838109 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838223 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838330 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838454 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.852419 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838455 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838541 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838703 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838719 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838858 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838931 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838961 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.838958 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839088 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839269 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839613 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839835 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839839 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.839933 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:30:51.339909881 +0000 UTC m=+20.543253298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.839918 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840143 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840353 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840375 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840443 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840603 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840659 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840716 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840827 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.840966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841237 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841464 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841639 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841696 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.841918 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842127 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842127 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842175 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842352 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842467 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842646 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842718 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842881 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842916 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.842978 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.846174 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.846259 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.846493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.846922 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.847061 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.847341 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.847379 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.847767 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.848107 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.848854 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.849152 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.849410 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.849645 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.849740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.849802 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.850134 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.850360 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.851147 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.851456 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.851535 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.851822 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.851991 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.855404 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.852587 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.858507 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.858666 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.859172 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.859401 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.859603 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.859947 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.860162 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.860434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.860505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.860576 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.860665 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861439 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861490 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861669 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.861825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.862257 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.862522 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.862583 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.862677 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.863630 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.863704 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.863955 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864114 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864251 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864305 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864468 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864612 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.864618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.865026 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.865225 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.867672 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.868176 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.868530 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.869910 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.870013 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:51.369985175 +0000 UTC m=+20.573328602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.870545 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.870876 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.871012 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:51.370992649 +0000 UTC m=+20.574336066 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.871862 4959 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.872090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.873815 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.881903 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.882296 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.883505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.890652 4959 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.891074 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.891240 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.891370 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.891574 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.891636 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:51.391596995 +0000 UTC m=+20.594940412 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.891530 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.897546 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.897919 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.898160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.901245 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.901441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.901486 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.901646 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.902052 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.902157 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.902239 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.902464 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.902791 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.904387 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.905316 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.906004 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.906431 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.907257 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.910812 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.911369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.913325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.913337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.913689 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.914909 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.914956 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.914974 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:50 crc kubenswrapper[4959]: E1003 13:30:50.915048 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:51.415022518 +0000 UTC m=+20.618366015 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.915878 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.916670 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.925997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.931612 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.934380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.942872 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.942946 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943048 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943084 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943101 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943112 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943122 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943132 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943144 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943153 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943162 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943174 4959 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943183 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943209 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943218 4959 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943230 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943241 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943250 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943259 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943269 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943279 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943287 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943291 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943321 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943333 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943348 4959 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943359 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943368 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943378 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943392 4959 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943403 4959 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943413 4959 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943422 4959 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943435 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943445 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943456 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943465 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943497 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943508 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943520 4959 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943536 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943551 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943563 4959 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943576 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943592 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943604 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943645 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943655 4959 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943676 4959 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943688 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943700 4959 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943717 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943728 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943738 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943747 4959 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943759 4959 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943772 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943799 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943811 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943825 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943834 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943843 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943853 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943868 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943879 4959 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943891 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943907 4959 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943919 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943930 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943939 4959 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943954 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943966 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943977 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.943988 4959 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944005 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944017 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944028 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944040 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944056 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944068 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944080 4959 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944096 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944107 4959 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944118 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944130 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944146 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944158 4959 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944172 4959 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944185 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944215 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944227 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944238 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944253 4959 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944281 4959 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944292 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944303 4959 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944318 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944330 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944341 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944353 4959 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944370 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944381 4959 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944392 4959 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944404 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944419 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944430 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944440 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944454 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944465 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944478 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944490 4959 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944505 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944516 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944528 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944541 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944556 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944567 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944578 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944588 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944603 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944619 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944630 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944646 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944657 4959 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944668 4959 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944681 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944952 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944970 4959 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944985 4959 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.944996 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945013 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945024 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945035 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945049 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945062 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945074 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945086 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945101 4959 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945114 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945125 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945136 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945153 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945165 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945176 4959 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945211 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945226 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945235 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945245 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945256 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945266 4959 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945277 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945287 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945299 4959 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945309 4959 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945319 4959 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945327 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945340 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945349 4959 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945358 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945370 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945379 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945389 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945398 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945411 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945421 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945429 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945440 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945452 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945461 4959 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945471 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945480 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945492 4959 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945501 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945512 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945522 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945532 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945541 4959 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945551 4959 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:50 crc kubenswrapper[4959]: I1003 13:30:50.945563 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.147860 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.156670 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.164433 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.165007 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:51 crc kubenswrapper[4959]: W1003 13:30:51.166725 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-0aa3e8654ee01820419c44715757a2c91e5e30975b4400033a6180b5b1c09323 WatchSource:0}: Error finding container 0aa3e8654ee01820419c44715757a2c91e5e30975b4400033a6180b5b1c09323: Status 404 returned error can't find the container with id 0aa3e8654ee01820419c44715757a2c91e5e30975b4400033a6180b5b1c09323 Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.172920 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:51 crc kubenswrapper[4959]: W1003 13:30:51.175486 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-97d51a24e2526f6589956f747bbe082cfbf63e341224f230e104691af1080812 WatchSource:0}: Error finding container 97d51a24e2526f6589956f747bbe082cfbf63e341224f230e104691af1080812: Status 404 returned error can't find the container with id 97d51a24e2526f6589956f747bbe082cfbf63e341224f230e104691af1080812 Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.190525 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.249041 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.249076 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.249089 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.350223 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.350398 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:30:52.35037999 +0000 UTC m=+21.553723407 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.451524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.451564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.451584 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.451604 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451706 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451762 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:52.451747868 +0000 UTC m=+21.655091285 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451773 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451805 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451799 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451896 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:52.451870521 +0000 UTC m=+21.655213988 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451819 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.451988 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:52.451966743 +0000 UTC m=+21.655310210 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.452056 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.452076 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.452115 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.452161 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:52.452150827 +0000 UTC m=+21.655494304 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.644802 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4dsjq"] Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.645172 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zskn9"] Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.645376 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.645654 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.646798 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-tfz2x"] Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.647058 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.647986 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.648489 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-v679m"] Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.648524 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.648787 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.648948 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.649065 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.649257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.649533 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.650212 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.651224 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.651592 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.651832 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.652047 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.652094 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.652969 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.653037 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.656862 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.658679 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.670562 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.684813 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.684835 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.684985 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:51 crc kubenswrapper[4959]: E1003 13:30:51.685048 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.685533 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.689419 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.690381 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.691971 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.692651 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.693738 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.694282 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.694987 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.696008 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.696577 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.696803 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.697968 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.698546 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.700109 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.703833 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.704837 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.706066 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.706649 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.707899 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.708800 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.709687 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.711917 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.712807 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.713567 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.714533 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.715543 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.716731 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.717704 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.718377 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.719456 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.719959 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.721024 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.721698 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.722229 4959 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.722784 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.723226 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.724574 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.725322 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.726683 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.728248 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.728980 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.729982 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.730682 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.731972 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.732624 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.733707 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.734481 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.735604 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.736492 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.736949 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.737611 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.738259 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.739649 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.740305 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.741674 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.742344 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.743238 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.744492 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.745158 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.750708 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-multus\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754400 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-os-release\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754632 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-os-release\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754730 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-system-cni-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-system-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.754985 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-netns\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755070 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-multus-certs\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755170 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-binary-copy\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-kubelet\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755389 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-multus-daemon-config\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755503 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4knxj\" (UniqueName: \"kubernetes.io/projected/cc425287-c256-4549-80ed-62c460cbe88e-kube-api-access-4knxj\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755647 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-etc-kubernetes\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgmc\" (UniqueName: \"kubernetes.io/projected/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-kube-api-access-plgmc\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755716 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-k8s-cni-cncf-io\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-hostroot\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755763 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wcdz\" (UniqueName: \"kubernetes.io/projected/117b68eb-14c1-456d-8c37-6259f7328587-kube-api-access-8wcdz\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755778 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-cni-binary-copy\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755798 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-rootfs\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755820 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-proxy-tls\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755849 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-cnibin\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755931 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/416afb5d-c0cb-4285-bd76-d7419c63cd54-hosts-file\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755957 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-bin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-cnibin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.755992 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-conf-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.756012 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2285t\" (UniqueName: \"kubernetes.io/projected/416afb5d-c0cb-4285-bd76-d7419c63cd54-kube-api-access-2285t\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.756028 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-socket-dir-parent\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.763475 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.774607 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.795746 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.800169 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"85928f3419d516cfe233e96a2f2cdb94db058a8f89ab2169cd92366aa955585e"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.801787 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.801848 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.801861 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"97d51a24e2526f6589956f747bbe082cfbf63e341224f230e104691af1080812"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.802849 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.802899 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0aa3e8654ee01820419c44715757a2c91e5e30975b4400033a6180b5b1c09323"} Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.833724 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.850044 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857309 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-cni-binary-copy\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-rootfs\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-proxy-tls\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857401 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857425 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857454 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-cnibin\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/416afb5d-c0cb-4285-bd76-d7419c63cd54-hosts-file\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857505 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-bin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857522 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-conf-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2285t\" (UniqueName: \"kubernetes.io/projected/416afb5d-c0cb-4285-bd76-d7419c63cd54-kube-api-access-2285t\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857560 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-cnibin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857578 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-socket-dir-parent\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857599 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-multus\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857618 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-os-release\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857637 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-os-release\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857680 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857705 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-system-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-netns\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-multus-certs\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-system-cni-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-kubelet\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857829 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-multus-daemon-config\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857852 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-binary-copy\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857936 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4knxj\" (UniqueName: \"kubernetes.io/projected/cc425287-c256-4549-80ed-62c460cbe88e-kube-api-access-4knxj\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857965 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgmc\" (UniqueName: \"kubernetes.io/projected/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-kube-api-access-plgmc\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.857987 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-k8s-cni-cncf-io\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858009 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-etc-kubernetes\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-hostroot\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858071 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wcdz\" (UniqueName: \"kubernetes.io/projected/117b68eb-14c1-456d-8c37-6259f7328587-kube-api-access-8wcdz\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858290 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-rootfs\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858368 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-os-release\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858774 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-netns\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858780 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-os-release\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.858998 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-cni-binary-copy\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859057 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-system-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859236 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-kubelet\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859288 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-cnibin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859326 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-etc-kubernetes\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859360 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-cni-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-cnibin\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859378 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-multus-certs\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-socket-dir-parent\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859395 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-multus\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859413 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-var-lib-cni-bin\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859411 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-hostroot\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859408 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-host-run-k8s-cni-cncf-io\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859417 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/416afb5d-c0cb-4285-bd76-d7419c63cd54-hosts-file\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859418 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-system-cni-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/117b68eb-14c1-456d-8c37-6259f7328587-multus-conf-dir\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.859733 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cc425287-c256-4549-80ed-62c460cbe88e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.860083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-binary-copy\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.860104 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/cc425287-c256-4549-80ed-62c460cbe88e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.860075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/117b68eb-14c1-456d-8c37-6259f7328587-multus-daemon-config\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.860659 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-mcd-auth-proxy-config\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.863706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-proxy-tls\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.867454 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.879740 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2285t\" (UniqueName: \"kubernetes.io/projected/416afb5d-c0cb-4285-bd76-d7419c63cd54-kube-api-access-2285t\") pod \"node-resolver-zskn9\" (UID: \"416afb5d-c0cb-4285-bd76-d7419c63cd54\") " pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.885685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgmc\" (UniqueName: \"kubernetes.io/projected/6639a7dd-61b2-4eb7-bf37-2ccbe2487dde-kube-api-access-plgmc\") pod \"machine-config-daemon-4dsjq\" (UID: \"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\") " pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.888326 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wcdz\" (UniqueName: \"kubernetes.io/projected/117b68eb-14c1-456d-8c37-6259f7328587-kube-api-access-8wcdz\") pod \"multus-tfz2x\" (UID: \"117b68eb-14c1-456d-8c37-6259f7328587\") " pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.894805 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4knxj\" (UniqueName: \"kubernetes.io/projected/cc425287-c256-4549-80ed-62c460cbe88e-kube-api-access-4knxj\") pod \"multus-additional-cni-plugins-v679m\" (UID: \"cc425287-c256-4549-80ed-62c460cbe88e\") " pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.901218 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.912095 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.924852 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.934400 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.945141 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.956961 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.959291 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.966866 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zskn9" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.969118 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.973284 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tfz2x" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.980693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v679m" Oct 03 13:30:51 crc kubenswrapper[4959]: W1003 13:30:51.987476 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod117b68eb_14c1_456d_8c37_6259f7328587.slice/crio-cc9745d4b7a5bad1bcd793c9f1e3b8d48e74d5f0b455706cde23a3aad8a19f2d WatchSource:0}: Error finding container cc9745d4b7a5bad1bcd793c9f1e3b8d48e74d5f0b455706cde23a3aad8a19f2d: Status 404 returned error can't find the container with id cc9745d4b7a5bad1bcd793c9f1e3b8d48e74d5f0b455706cde23a3aad8a19f2d Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.988748 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:51 crc kubenswrapper[4959]: I1003 13:30:51.999442 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.010715 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.016216 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xqbk4"] Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.016968 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019499 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019551 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019612 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019518 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019708 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019748 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.019951 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.021635 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.029178 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.044849 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.063405 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.076626 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.096928 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.118410 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.154909 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164347 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164385 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164405 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164428 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164443 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tjvp\" (UniqueName: \"kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164462 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164479 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164551 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164604 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164642 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164712 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164738 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164773 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164801 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164835 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164889 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164913 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.164933 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.180338 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.199273 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.220651 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.233846 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.248482 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.261032 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266405 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266451 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tjvp\" (UniqueName: \"kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266486 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266508 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266520 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266529 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266577 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266616 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266648 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266670 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266687 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266710 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266743 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266800 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266819 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266889 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266893 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266897 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.266986 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267022 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267052 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267034 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267088 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267100 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267106 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267134 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267177 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267669 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.267826 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.272644 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.275907 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.283878 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tjvp\" (UniqueName: \"kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp\") pod \"ovnkube-node-xqbk4\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.291122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.308906 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.321896 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.336472 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.353176 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.357118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.368215 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.368460 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:30:54.368438878 +0000 UTC m=+23.571782295 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.369688 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.382966 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.398326 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.416763 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.436773 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.455253 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.468285 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.468739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.468773 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.468800 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.468821 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.468947 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.468966 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.468994 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469002 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469060 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:54.469036948 +0000 UTC m=+23.672380365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469084 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:54.469077568 +0000 UTC m=+23.672420985 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469118 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469131 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469149 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469147 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469228 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:54.469217282 +0000 UTC m=+23.672560699 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.469306 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:54.469272863 +0000 UTC m=+23.672616440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.483869 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.514271 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.538926 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.685360 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:52 crc kubenswrapper[4959]: E1003 13:30:52.685483 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.807948 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zskn9" event={"ID":"416afb5d-c0cb-4285-bd76-d7419c63cd54","Type":"ContainerStarted","Data":"e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.808012 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zskn9" event={"ID":"416afb5d-c0cb-4285-bd76-d7419c63cd54","Type":"ContainerStarted","Data":"2d8f338df9c113c14af2e1e4375cbf3458961eb5d84e8acd1381ac2de08a4c63"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.809876 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.809916 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.809931 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"3e0801eeeb5db19ba1bb49261a63edfde6541d8ba07a8612e8f785628a76c8b5"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.811936 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122" exitCode=0 Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.812028 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.812065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerStarted","Data":"c2a5340463effd4eb239404b84ad256ba946b3dfd11daf80d8376ae90e0f2afa"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.815048 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" exitCode=0 Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.815123 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.815156 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"4869ec5f0a19823454df7cbb763a45e2ceae6d03714f2576a9cabe43b08a7648"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.817702 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerStarted","Data":"68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.817770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerStarted","Data":"cc9745d4b7a5bad1bcd793c9f1e3b8d48e74d5f0b455706cde23a3aad8a19f2d"} Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.834072 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.855853 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.869168 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.902694 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.960728 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:52 crc kubenswrapper[4959]: I1003 13:30:52.981964 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.000862 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:52Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.022774 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.036394 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.050893 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.063496 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.091653 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.108734 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.127106 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.149966 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.161483 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.178137 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.196184 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.223532 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.245182 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.258690 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.276543 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.306245 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.349810 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.688877 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:53 crc kubenswrapper[4959]: E1003 13:30:53.689282 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.689356 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:53 crc kubenswrapper[4959]: E1003 13:30:53.689415 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.823923 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d" exitCode=0 Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.824001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.825792 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.831109 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.831166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.831185 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.831224 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.838362 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.852395 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.871775 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.888753 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.905726 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.920316 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.935386 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.948807 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.964034 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.979049 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:53 crc kubenswrapper[4959]: I1003 13:30:53.997634 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:53Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.022324 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.036974 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.051505 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.064847 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.082219 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.105165 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.118111 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.128215 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.140089 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.143148 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.146652 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.166397 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.204430 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.244906 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.278949 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.283803 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.285977 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.303620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.345473 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.384433 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.391943 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.392143 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:30:58.392113083 +0000 UTC m=+27.595456520 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.423784 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.463033 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.492709 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.492783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.492841 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.492882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.492901 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.492930 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.492934 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.492971 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.492944 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493028 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:58.49300522 +0000 UTC m=+27.696348637 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493044 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493071 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493082 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493048 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:58.49303896 +0000 UTC m=+27.696382497 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493110 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:58.493101792 +0000 UTC m=+27.696445339 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.493140 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:30:58.493124392 +0000 UTC m=+27.696467809 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.508736 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.548129 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.582757 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.647030 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.675980 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.684377 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:54 crc kubenswrapper[4959]: E1003 13:30:54.684481 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.705931 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.723822 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-9gj84"] Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.724275 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.743476 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.756263 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.775904 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.794471 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22c9cebf-ffbf-4281-9200-58029e483f64-host\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.794515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczm5\" (UniqueName: \"kubernetes.io/projected/22c9cebf-ffbf-4281-9200-58029e483f64-kube-api-access-pczm5\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.794538 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22c9cebf-ffbf-4281-9200-58029e483f64-serviceca\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.796394 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.815699 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.837991 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4" exitCode=0 Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.838090 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4"} Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.842283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.842323 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.864474 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.895369 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22c9cebf-ffbf-4281-9200-58029e483f64-host\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.895417 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczm5\" (UniqueName: \"kubernetes.io/projected/22c9cebf-ffbf-4281-9200-58029e483f64-kube-api-access-pczm5\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.895441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22c9cebf-ffbf-4281-9200-58029e483f64-serviceca\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.895461 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22c9cebf-ffbf-4281-9200-58029e483f64-host\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.896402 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22c9cebf-ffbf-4281-9200-58029e483f64-serviceca\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.904019 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.932114 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczm5\" (UniqueName: \"kubernetes.io/projected/22c9cebf-ffbf-4281-9200-58029e483f64-kube-api-access-pczm5\") pod \"node-ca-9gj84\" (UID: \"22c9cebf-ffbf-4281-9200-58029e483f64\") " pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:54 crc kubenswrapper[4959]: I1003 13:30:54.967317 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:54Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.010701 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.034967 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9gj84" Oct 03 13:30:55 crc kubenswrapper[4959]: W1003 13:30:55.047595 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22c9cebf_ffbf_4281_9200_58029e483f64.slice/crio-9c1f556b1036a41bcc3499a646f3d0fdd45f14654289c819ea66e226ae474814 WatchSource:0}: Error finding container 9c1f556b1036a41bcc3499a646f3d0fdd45f14654289c819ea66e226ae474814: Status 404 returned error can't find the container with id 9c1f556b1036a41bcc3499a646f3d0fdd45f14654289c819ea66e226ae474814 Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.051735 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.085603 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.123710 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.162917 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.201624 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.243858 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.285516 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.323743 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.365933 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.406016 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.446486 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.484271 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.534336 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.564527 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.603975 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.684747 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:55 crc kubenswrapper[4959]: E1003 13:30:55.684939 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.685115 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:55 crc kubenswrapper[4959]: E1003 13:30:55.685371 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.850666 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb" exitCode=0 Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.850754 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb"} Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.852642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9gj84" event={"ID":"22c9cebf-ffbf-4281-9200-58029e483f64","Type":"ContainerStarted","Data":"0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3"} Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.852782 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9gj84" event={"ID":"22c9cebf-ffbf-4281-9200-58029e483f64","Type":"ContainerStarted","Data":"9c1f556b1036a41bcc3499a646f3d0fdd45f14654289c819ea66e226ae474814"} Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.875919 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.889276 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.899789 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.917962 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.934351 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.951028 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.967357 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.980826 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:55 crc kubenswrapper[4959]: I1003 13:30:55.995305 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:55Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.005319 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.052038 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.084022 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.128654 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.163380 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.213556 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.254126 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.285967 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.322470 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.370183 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.403280 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.443295 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.486258 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.529668 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.564270 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.605937 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.645986 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.683927 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.684949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:56 crc kubenswrapper[4959]: E1003 13:30:56.685083 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.728996 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.774574 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.810575 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.860268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.864210 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e" exitCode=0 Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.864250 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e"} Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.881496 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.899445 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.925828 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:56 crc kubenswrapper[4959]: I1003 13:30:56.962623 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:56Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.009074 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.045096 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.083331 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.131705 4959 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.133724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.133754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.133764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.133862 4959 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.136161 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.176047 4959 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.176398 4959 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.177503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.177618 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.177713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.177798 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.177882 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.193175 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.197297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.197451 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.197563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.197690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.197775 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.203237 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.208625 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.212523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.212570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.212580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.212597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.212609 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.223988 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.227250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.227280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.227290 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.227304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.227314 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.240343 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243692 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.243666 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.255147 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.255470 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.256985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.257097 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.257164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.257258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.257324 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.282021 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.323467 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.359774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.359810 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.359819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.359832 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.359841 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.368312 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.407990 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.445145 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.461974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.462217 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.462280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.462341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.462427 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.565176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.565281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.565295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.565317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.565333 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.668149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.668208 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.668219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.668234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.668244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.685364 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.685516 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.685556 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:57 crc kubenswrapper[4959]: E1003 13:30:57.685811 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.771496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.771543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.771552 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.771570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.771583 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.872691 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc425287-c256-4549-80ed-62c460cbe88e" containerID="45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e" exitCode=0 Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.872756 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerDied","Data":"45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.873713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.873752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.873763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.873782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.873793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.891865 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.909806 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.922798 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.934033 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.949269 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.966365 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.976114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.976150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.976160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.976174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.976182 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:57Z","lastTransitionTime":"2025-10-03T13:30:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.978851 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:57 crc kubenswrapper[4959]: I1003 13:30:57.995956 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:57Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.007887 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.019294 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.028216 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.051532 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.066878 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.078259 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.078297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.078305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.078319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.078329 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.082042 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.102969 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.180420 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.180479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.180496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.180518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.180533 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.283715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.283764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.283777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.283794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.283807 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.386377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.386699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.386715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.386739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.386755 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.432752 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.433066 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.433012948 +0000 UTC m=+35.636356395 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.489631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.489677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.489687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.489702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.489712 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.534398 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.534454 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.534496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.534526 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534594 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534647 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534711 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.534686213 +0000 UTC m=+35.738029690 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534736 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.534728964 +0000 UTC m=+35.738072371 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534767 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534821 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534838 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534789 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534903 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.534883638 +0000 UTC m=+35.738227065 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534916 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534935 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.534985 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.53496773 +0000 UTC m=+35.738311147 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.592569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.592630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.592642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.592661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.592672 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.686121 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:30:58 crc kubenswrapper[4959]: E1003 13:30:58.686352 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.695562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.695621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.695637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.695664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.695681 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.798784 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.798829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.798838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.798858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.798869 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.884106 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.884429 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.888892 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" event={"ID":"cc425287-c256-4549-80ed-62c460cbe88e","Type":"ContainerStarted","Data":"30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.899474 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.901819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.901860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.901870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.901888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.901901 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:58Z","lastTransitionTime":"2025-10-03T13:30:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.907963 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.912934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.933521 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.953122 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.969565 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.981583 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:58 crc kubenswrapper[4959]: I1003 13:30:58.994306 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004365 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.004686 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.013411 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.026748 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.045570 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.062561 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.076165 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.087367 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.097351 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.106258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.106302 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.106315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.106335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.106349 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.119324 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.133603 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.153112 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.172539 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.186882 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.200087 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.208718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.208772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.208785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.208805 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.208821 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.216847 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.245942 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.258831 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.273138 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.282080 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.288497 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.301055 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.310984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.311038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.311055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.311078 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.311095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.314914 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.332716 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.349945 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.365815 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.380803 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.395303 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.410944 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.414231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.414273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.414282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.414301 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.414315 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.446034 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.484520 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.517079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.517125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.517135 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.517152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.517163 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.527187 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.570896 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.606877 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.620174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.620231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.620241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.620261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.620274 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.644981 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.684609 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.684727 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:30:59 crc kubenswrapper[4959]: E1003 13:30:59.684774 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:30:59 crc kubenswrapper[4959]: E1003 13:30:59.684946 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.686747 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.723483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.723541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.723555 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.723574 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.723586 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.726473 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.764349 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.805917 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.825858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.825902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.825914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.825929 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.825938 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.850912 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.892667 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.893132 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.920774 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.928253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.928308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.928321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.928339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.928354 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:30:59Z","lastTransitionTime":"2025-10-03T13:30:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.938182 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.958218 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:30:59 crc kubenswrapper[4959]: I1003 13:30:59.981799 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:30:59Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.008357 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.031172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.031262 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.031275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.031292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.031307 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.045947 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.088396 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.125789 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.133674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.133796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.133887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.133992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.134073 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.166043 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.206271 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.236992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.237036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.237046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.237070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.237084 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.243813 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.290456 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.325865 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.339469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.339514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.339524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.339543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.339556 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.363861 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.405667 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.442889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.442931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.442943 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.442960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.442975 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.443309 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:00Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.546850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.546906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.546999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.547025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.547043 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.657913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.657960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.657973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.657993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.658005 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.684835 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:00 crc kubenswrapper[4959]: E1003 13:31:00.685065 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.761706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.761761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.761774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.761794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.761806 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.863791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.863856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.863869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.863887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.863900 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.896384 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.965913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.965940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.965948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.965961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:00 crc kubenswrapper[4959]: I1003 13:31:00.965970 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:00Z","lastTransitionTime":"2025-10-03T13:31:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.069534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.069582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.069596 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.069616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.069628 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.172289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.172359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.172372 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.172395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.172412 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.274936 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.274990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.275004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.275025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.275038 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.378394 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.378469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.378482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.378503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.378529 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.481932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.481974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.481985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.482003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.482015 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.584785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.585363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.585374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.585389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.585400 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.685599 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.685641 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:01 crc kubenswrapper[4959]: E1003 13:31:01.685853 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:01 crc kubenswrapper[4959]: E1003 13:31:01.686010 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.691323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.691395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.691410 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.691437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.691463 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.705039 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.718774 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.738218 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.757066 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.769801 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.780761 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.789559 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.804648 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.807736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.807777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.807788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.807806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.807817 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.824635 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.837228 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.851476 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.867743 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.883548 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.897816 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.902752 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/0.log" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.906044 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9" exitCode=1 Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.906121 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.906827 4959 scope.go:117] "RemoveContainer" containerID="ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.909927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.910812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.910894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.911032 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.911100 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:01Z","lastTransitionTime":"2025-10-03T13:31:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.920827 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.934000 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.947056 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.973235 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:01Z\\\",\\\"message\\\":\\\" 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 13:31:01.656227 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:01.656269 6245 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1003 13:31:01.656276 6245 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1003 13:31:01.656291 6245 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 13:31:01.656295 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:01.656308 6245 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1003 13:31:01.656323 6245 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1003 13:31:01.656327 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 13:31:01.656333 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 13:31:01.656342 6245 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:01.656356 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:01.656366 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 13:31:01.656399 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:01.656399 6245 factory.go:656] Stopping watch factory\\\\nI1003 13:31:01.656417 6245 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:01 crc kubenswrapper[4959]: I1003 13:31:01.991674 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.010065 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.015449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.015517 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.015622 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.015671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.015711 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.026098 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.039876 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.056031 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.067088 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.077896 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.096651 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.110640 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.119295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.119336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.119348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.119365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.119377 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.123455 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.136731 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.147666 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.221972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.222024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.222036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.222057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.222069 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.324124 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.324171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.324182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.324226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.324240 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.426411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.426461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.426473 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.426489 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.426500 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.532365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.532410 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.532424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.532446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.532462 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.634906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.635018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.635034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.635049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.635062 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.684857 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:02 crc kubenswrapper[4959]: E1003 13:31:02.685000 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.738626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.738672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.738688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.738706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.738717 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.837266 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp"] Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.837665 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.840128 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.840460 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.841299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.841328 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.841339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.841354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.841365 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.862214 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.878736 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.890425 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.900995 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.910029 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/1.log" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.910749 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/0.log" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.913115 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" exitCode=1 Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.913166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.913244 4959 scope.go:117] "RemoveContainer" containerID="ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.914085 4959 scope.go:117] "RemoveContainer" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" Oct 03 13:31:02 crc kubenswrapper[4959]: E1003 13:31:02.914396 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.915523 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.917881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.917931 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.917952 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rnvk\" (UniqueName: \"kubernetes.io/projected/da45dabb-f9ff-428b-904c-f45a1cf701a5-kube-api-access-9rnvk\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.917981 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.925635 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.936676 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.943802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.943843 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.943852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.943866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.943877 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:02Z","lastTransitionTime":"2025-10-03T13:31:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.946784 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.966727 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:01Z\\\",\\\"message\\\":\\\" 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 13:31:01.656227 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:01.656269 6245 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1003 13:31:01.656276 6245 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1003 13:31:01.656291 6245 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 13:31:01.656295 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:01.656308 6245 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1003 13:31:01.656323 6245 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1003 13:31:01.656327 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 13:31:01.656333 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 13:31:01.656342 6245 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:01.656356 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:01.656366 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 13:31:01.656399 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:01.656399 6245 factory.go:656] Stopping watch factory\\\\nI1003 13:31:01.656417 6245 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.979695 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:02 crc kubenswrapper[4959]: I1003 13:31:02.990890 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.006264 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.019050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.019367 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.019477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.019588 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rnvk\" (UniqueName: \"kubernetes.io/projected/da45dabb-f9ff-428b-904c-f45a1cf701a5-kube-api-access-9rnvk\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.020141 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.020428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.022762 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.026628 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da45dabb-f9ff-428b-904c-f45a1cf701a5-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.058627 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rnvk\" (UniqueName: \"kubernetes.io/projected/da45dabb-f9ff-428b-904c-f45a1cf701a5-kube-api-access-9rnvk\") pod \"ovnkube-control-plane-749d76644c-mxxqp\" (UID: \"da45dabb-f9ff-428b-904c-f45a1cf701a5\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.058939 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.059441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.059471 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.059483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.059500 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.059512 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.079699 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.105695 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.117130 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.129907 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.152575 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.152929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.163419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.163495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.163509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.163530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.163546 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.169137 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.181775 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.194949 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.205534 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.217947 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.243569 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff858022e05fc8c398f8ef2d6f5688b6c7f12e94945edca97b1bd6478ecbd2a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:01Z\\\",\\\"message\\\":\\\" 6245 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 13:31:01.656227 6245 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:01.656269 6245 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1003 13:31:01.656276 6245 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1003 13:31:01.656291 6245 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 13:31:01.656295 6245 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:01.656308 6245 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1003 13:31:01.656323 6245 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1003 13:31:01.656327 6245 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 13:31:01.656333 6245 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 13:31:01.656342 6245 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:01.656356 6245 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:01.656366 6245 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 13:31:01.656399 6245 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:01.656399 6245 factory.go:656] Stopping watch factory\\\\nI1003 13:31:01.656417 6245 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.252304 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.266415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.266457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.266468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.266483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.266492 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.267009 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.280643 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.291707 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.302212 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.312118 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.321621 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.368984 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.369016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.369025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.369037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.369046 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.471469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.471510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.471520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.471536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.471545 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.574951 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.575053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.575130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.575241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.575278 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.678377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.678456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.678474 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.678498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.678515 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.684791 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.684935 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:03 crc kubenswrapper[4959]: E1003 13:31:03.685139 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:03 crc kubenswrapper[4959]: E1003 13:31:03.685363 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.781332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.781407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.781430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.781461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.781485 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.885459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.885798 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.885812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.885830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.885843 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.918295 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/1.log" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.923505 4959 scope.go:117] "RemoveContainer" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.923584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" event={"ID":"da45dabb-f9ff-428b-904c-f45a1cf701a5","Type":"ContainerStarted","Data":"5f8b4ac1397985e094caea71fa97eb5873fccb8891cb68d66583ee258c84aaf1"} Oct 03 13:31:03 crc kubenswrapper[4959]: E1003 13:31:03.923825 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.936758 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.957132 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.983082 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:03Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.988062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.988112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.988128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.988148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:03 crc kubenswrapper[4959]: I1003 13:31:03.988162 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:03Z","lastTransitionTime":"2025-10-03T13:31:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.004736 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.023509 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.044937 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.062776 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.078523 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.092477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.092549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.092562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.092581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.092594 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.103942 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.117006 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.136024 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.152435 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.171237 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.183100 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.194951 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.195013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.195024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.195041 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.195053 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.197290 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.209344 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.298071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.298111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.298121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.298137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.298148 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.401059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.401114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.401125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.401141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.401152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.504059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.504125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.504137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.504153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.504164 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.607028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.607080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.607091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.607110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.607123 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.685422 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:04 crc kubenswrapper[4959]: E1003 13:31:04.685598 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.708953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.709000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.709011 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.709027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.709040 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.812053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.812103 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.812111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.812131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.812143 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.917774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.917859 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.917879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.917905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.917922 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:04Z","lastTransitionTime":"2025-10-03T13:31:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.928939 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" event={"ID":"da45dabb-f9ff-428b-904c-f45a1cf701a5","Type":"ContainerStarted","Data":"09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.928982 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" event={"ID":"da45dabb-f9ff-428b-904c-f45a1cf701a5","Type":"ContainerStarted","Data":"26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4"} Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.949682 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.956652 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.958956 4959 scope.go:117] "RemoveContainer" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" Oct 03 13:31:04 crc kubenswrapper[4959]: E1003 13:31:04.959317 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.967323 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.980388 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:04 crc kubenswrapper[4959]: I1003 13:31:04.992555 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:04Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.002385 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.015365 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.019927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.019957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.019966 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.019982 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.019992 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.030092 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.041978 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.055954 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.069302 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.084428 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.086727 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gf4tr"] Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.087303 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.087378 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.097438 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.112777 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.122125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.122174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.122185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.122228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.122241 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.131934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.144444 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.156914 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.172404 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.198862 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.217654 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.224838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.224870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.224881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.224899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.224909 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.234737 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.243929 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.244017 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4shq\" (UniqueName: \"kubernetes.io/projected/e3a679d1-7412-4843-998e-8d1c109bf7ec-kube-api-access-k4shq\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.249495 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.264811 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.281435 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.294924 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.305749 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.321814 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.326987 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.327033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.327045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.327064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.327076 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.337602 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.345063 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.345542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4shq\" (UniqueName: \"kubernetes.io/projected/e3a679d1-7412-4843-998e-8d1c109bf7ec-kube-api-access-k4shq\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.345761 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.345937 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:05.845887572 +0000 UTC m=+35.049230989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.354263 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.365421 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4shq\" (UniqueName: \"kubernetes.io/projected/e3a679d1-7412-4843-998e-8d1c109bf7ec-kube-api-access-k4shq\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.365892 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.378548 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.390775 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.400124 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.410501 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:05Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.429658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.429710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.429722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.429742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.429755 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.532646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.532701 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.532711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.532728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.532740 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.635480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.635749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.635903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.636014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.636131 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.685241 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.685241 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.685423 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.685472 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.739303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.739374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.739388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.739406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.739419 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.842181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.842301 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.842321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.842351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.842372 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.849942 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.850165 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:05 crc kubenswrapper[4959]: E1003 13:31:05.850353 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:06.850315155 +0000 UTC m=+36.053658612 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.944744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.944804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.944813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.944830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:05 crc kubenswrapper[4959]: I1003 13:31:05.944843 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:05Z","lastTransitionTime":"2025-10-03T13:31:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.047940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.048043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.048059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.048079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.048095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.151224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.151269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.151279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.151299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.151312 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.254576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.254627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.254639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.254655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.254666 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.358004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.358412 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.358503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.358598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.358680 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.457130 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.457361 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:31:22.457325895 +0000 UTC m=+51.660669312 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.462581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.462646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.462659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.462680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.462694 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.559655 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.559778 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.559796 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.559675 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.559890 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:22.559848052 +0000 UTC m=+51.763191469 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.559915 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.559957 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.559979 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560128 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560150 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560163 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560245 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:22.560225571 +0000 UTC m=+51.763569178 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560326 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560353 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:22.560345693 +0000 UTC m=+51.763689340 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560385 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.560408 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:22.560400705 +0000 UTC m=+51.763744122 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.566249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.566300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.566314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.566334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.566348 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.668945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.669020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.669030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.669056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.669068 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.685370 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.685428 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.685537 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.685711 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.771692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.771756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.771765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.771781 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.771793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.863183 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.863419 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: E1003 13:31:06.863497 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:08.863473134 +0000 UTC m=+38.066816551 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.874602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.874645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.874655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.874677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.874690 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.977490 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.977557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.977577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.977601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:06 crc kubenswrapper[4959]: I1003 13:31:06.977621 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:06Z","lastTransitionTime":"2025-10-03T13:31:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.080440 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.080521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.080546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.080573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.080592 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.184764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.184848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.184877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.184913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.184938 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.288077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.288139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.288155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.288174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.288210 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.395171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.395300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.395333 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.395362 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.395392 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.498671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.498746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.498758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.498779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.498791 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.586674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.586744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.586756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.586780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.586792 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.602245 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:07Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.607565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.607621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.607634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.607654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.607673 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.621232 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:07Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.625998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.626060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.626072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.626091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.626105 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.639027 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:07Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.643555 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.643592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.643601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.643622 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.643636 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.657590 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:07Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.662244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.662286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.662297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.662314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.662327 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.674987 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:07Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.675162 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.677476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.677521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.677534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.677559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.677573 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.684653 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.684810 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.684856 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:07 crc kubenswrapper[4959]: E1003 13:31:07.684995 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.780253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.780296 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.780310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.780328 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.780342 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.882625 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.882674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.882683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.882705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.882721 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.990273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.990335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.990347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.990379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:07 crc kubenswrapper[4959]: I1003 13:31:07.990395 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:07Z","lastTransitionTime":"2025-10-03T13:31:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.093885 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.093960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.093983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.094014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.094042 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.198170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.198265 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.198286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.198311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.198324 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.301067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.301146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.301165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.301189 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.301252 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.403349 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.403391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.403399 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.403416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.403427 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.507108 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.507163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.507179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.507219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.507232 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.610287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.610366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.610388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.610420 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.610443 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.684514 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:08 crc kubenswrapper[4959]: E1003 13:31:08.684698 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.684514 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:08 crc kubenswrapper[4959]: E1003 13:31:08.684881 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.714024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.714076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.714096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.714119 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.714137 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.817309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.817350 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.817359 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.817375 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.817385 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.887162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:08 crc kubenswrapper[4959]: E1003 13:31:08.887430 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:08 crc kubenswrapper[4959]: E1003 13:31:08.887548 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:12.887521518 +0000 UTC m=+42.090865025 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.920348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.920401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.920417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.920436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:08 crc kubenswrapper[4959]: I1003 13:31:08.920449 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:08Z","lastTransitionTime":"2025-10-03T13:31:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.023693 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.023768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.023793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.023822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.023846 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.127094 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.127157 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.127166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.127182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.127206 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.229933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.230006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.230038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.230062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.230077 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.332416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.332512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.332528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.332551 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.332569 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.434960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.435021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.435039 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.435062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.435081 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.538067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.538127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.538137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.538155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.538166 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.640819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.640890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.640908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.640935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.640952 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.685587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.685590 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:09 crc kubenswrapper[4959]: E1003 13:31:09.685741 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:09 crc kubenswrapper[4959]: E1003 13:31:09.685838 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.743184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.743235 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.743244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.743258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.743268 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.846743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.846788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.846798 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.846817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.846830 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.949067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.949126 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.949135 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.949149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:09 crc kubenswrapper[4959]: I1003 13:31:09.949161 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:09Z","lastTransitionTime":"2025-10-03T13:31:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.051630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.051700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.051710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.051737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.051749 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.154947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.155014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.155025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.155046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.155059 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.258139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.258224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.258249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.258268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.258280 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.361277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.361342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.361365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.361395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.361419 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.463881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.463950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.463961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.463981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.464320 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.566340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.566425 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.566439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.566453 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.566463 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.671690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.671759 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.671779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.671814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.671831 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.684781 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.684910 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:10 crc kubenswrapper[4959]: E1003 13:31:10.684943 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:10 crc kubenswrapper[4959]: E1003 13:31:10.685109 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.774466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.774507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.774516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.774530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.774540 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.877342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.877805 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.878067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.878347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.878567 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.981443 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.981480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.981491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.981512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:10 crc kubenswrapper[4959]: I1003 13:31:10.981525 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:10Z","lastTransitionTime":"2025-10-03T13:31:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.084156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.084276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.084303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.084334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.084355 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.187613 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.187665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.187677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.187696 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.187710 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.299152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.299280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.299298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.299323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.299345 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.402437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.402488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.402499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.402518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.402529 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.505979 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.506039 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.506053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.506075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.506088 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.609385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.609445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.609464 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.609487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.609503 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.685487 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.685487 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:11 crc kubenswrapper[4959]: E1003 13:31:11.685746 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:11 crc kubenswrapper[4959]: E1003 13:31:11.685656 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.707294 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.711959 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.712006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.712024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.712047 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.712063 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.720955 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.737743 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.756374 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.775025 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.788886 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.804755 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.813649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.813695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.813708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.813727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.813743 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.824134 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.835815 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.847853 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.870829 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.890021 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.905405 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.916180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.916253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.916264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.916284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.916296 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:11Z","lastTransitionTime":"2025-10-03T13:31:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.919764 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.931644 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.945965 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:11 crc kubenswrapper[4959]: I1003 13:31:11.955551 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:11Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.019373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.019422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.019435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.019455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.019472 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.122865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.122911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.122923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.122939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.122949 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.226667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.226716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.226731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.226751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.226763 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.329253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.329313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.329326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.329345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.329359 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.432288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.432341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.432355 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.432378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.432393 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.536094 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.536275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.536295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.536328 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.536341 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.638687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.638748 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.638762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.638791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.638817 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.684417 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.684451 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:12 crc kubenswrapper[4959]: E1003 13:31:12.684695 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:12 crc kubenswrapper[4959]: E1003 13:31:12.684774 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.740967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.741004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.741012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.741026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.741034 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.843535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.843582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.843601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.843623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.843637 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.931839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:12 crc kubenswrapper[4959]: E1003 13:31:12.931989 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:12 crc kubenswrapper[4959]: E1003 13:31:12.932071 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:20.93205119 +0000 UTC m=+50.135394617 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.947498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.947548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.947567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.947588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:12 crc kubenswrapper[4959]: I1003 13:31:12.947600 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:12Z","lastTransitionTime":"2025-10-03T13:31:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.051239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.051285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.051297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.051316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.051329 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.154596 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.154698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.154721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.154752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.154775 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.257893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.257936 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.257945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.257961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.257972 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.364968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.365023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.365038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.365056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.365078 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.468006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.468069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.468083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.468105 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.468119 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.572056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.572114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.572123 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.572140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.572152 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.674910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.674972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.674989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.675010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.675026 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.685627 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.685660 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:13 crc kubenswrapper[4959]: E1003 13:31:13.685766 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:13 crc kubenswrapper[4959]: E1003 13:31:13.685942 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.778343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.778395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.778408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.778428 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.778441 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.882153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.882232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.882247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.882267 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.882279 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.985181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.985265 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.985275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.985294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:13 crc kubenswrapper[4959]: I1003 13:31:13.985306 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:13Z","lastTransitionTime":"2025-10-03T13:31:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.088388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.088468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.088479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.088498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.088510 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.191528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.191605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.191639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.191669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.191684 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.295175 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.295272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.295287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.295307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.295338 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.398307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.398398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.398414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.398438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.398452 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.501329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.501383 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.501398 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.501424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.501438 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.605838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.605902 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.605919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.605945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.605965 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.685293 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.685405 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:14 crc kubenswrapper[4959]: E1003 13:31:14.685510 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:14 crc kubenswrapper[4959]: E1003 13:31:14.685625 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.708697 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.708775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.708792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.708836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.708850 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.811908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.811960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.811973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.811995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.812008 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.915080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.915139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.915152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.915176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:14 crc kubenswrapper[4959]: I1003 13:31:14.915216 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:14Z","lastTransitionTime":"2025-10-03T13:31:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.017722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.017767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.017786 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.017806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.017820 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.120809 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.120872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.120889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.120917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.120953 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.224127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.224226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.224248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.224275 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.224288 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.327160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.327248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.327269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.327291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.327307 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.430385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.430470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.430504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.430534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.430554 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.533585 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.533645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.533661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.533681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.533860 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.637055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.637122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.637136 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.637158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.637171 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.685479 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.685479 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:15 crc kubenswrapper[4959]: E1003 13:31:15.685623 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:15 crc kubenswrapper[4959]: E1003 13:31:15.685722 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.741008 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.741074 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.741087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.741111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.741127 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.843849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.843919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.843930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.843952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.843966 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.946260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.946312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.946323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.946346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:15 crc kubenswrapper[4959]: I1003 13:31:15.946358 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:15Z","lastTransitionTime":"2025-10-03T13:31:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.048761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.048819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.048831 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.048849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.048862 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.151856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.151905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.151918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.151938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.151949 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.255249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.255298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.255309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.255330 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.255340 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.358747 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.358816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.358834 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.358858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.358874 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.462056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.462116 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.462133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.462167 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.462180 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.564917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.564990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.565006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.565029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.565046 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.668945 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.669003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.669014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.669038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.669056 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.684677 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.684761 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:16 crc kubenswrapper[4959]: E1003 13:31:16.684909 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:16 crc kubenswrapper[4959]: E1003 13:31:16.685026 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.771583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.771624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.771633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.771648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.771658 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.874185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.874304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.874324 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.874352 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.874374 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.976024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.976059 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.976068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.976082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:16 crc kubenswrapper[4959]: I1003 13:31:16.976093 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:16Z","lastTransitionTime":"2025-10-03T13:31:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.080373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.080478 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.080502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.080577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.080604 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.184163 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.184292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.184312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.184337 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.184356 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.287229 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.287289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.287303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.287328 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.287343 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.390688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.390746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.390757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.390776 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.390789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.493357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.493396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.493407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.493424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.493437 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.597298 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.597357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.597370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.597391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.597407 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.685444 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.685626 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.685775 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.686271 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.686410 4959 scope.go:117] "RemoveContainer" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.699822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.699854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.699862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.699877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.699886 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.804020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.804077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.804091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.804112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.804125 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.906104 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.906147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.906159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.906177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.906213 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.910452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.910503 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.910513 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.910529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.910540 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.923076 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:17Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.927651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.927789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.928105 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.928131 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.928144 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.943087 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:17Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.948523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.948579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.948590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.948610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.948625 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.965627 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:17Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.971058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.971108 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.971120 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.971140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.971155 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.973427 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/1.log" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.976972 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.977718 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:31:17 crc kubenswrapper[4959]: E1003 13:31:17.987697 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:17Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.992268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.992335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.992347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.992388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.992406 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:17Z","lastTransitionTime":"2025-10-03T13:31:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:17 crc kubenswrapper[4959]: I1003 13:31:17.997620 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:17Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: E1003 13:31:18.010242 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: E1003 13:31:18.010442 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.012919 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.012968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.012980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.013002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.013015 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.013628 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.028849 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.050802 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.075406 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.092324 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.115867 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.116559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.116635 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.116655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.116677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.116692 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.132172 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.145867 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.171686 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.189511 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.218846 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.220255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.220314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.220326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.220347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.220652 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.235997 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.255642 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.273742 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.288091 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.307760 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:18Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.325278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.325315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.325324 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.325340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.325349 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.428663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.428720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.428730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.428752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.428765 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.531514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.531562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.531575 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.531593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.531605 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.634681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.634728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.634737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.634758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.634770 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.685479 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.685508 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:18 crc kubenswrapper[4959]: E1003 13:31:18.685661 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:18 crc kubenswrapper[4959]: E1003 13:31:18.685813 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.737344 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.737395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.737406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.737429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.737443 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.840790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.840836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.840846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.840863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.840873 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.944118 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.944164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.944177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.944220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.944237 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:18Z","lastTransitionTime":"2025-10-03T13:31:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.983419 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/2.log" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.984536 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/1.log" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.988862 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" exitCode=1 Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.988924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9"} Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.989137 4959 scope.go:117] "RemoveContainer" containerID="cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa" Oct 03 13:31:18 crc kubenswrapper[4959]: I1003 13:31:18.989648 4959 scope.go:117] "RemoveContainer" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" Oct 03 13:31:18 crc kubenswrapper[4959]: E1003 13:31:18.989984 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.010041 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.028545 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.047554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.047599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.047612 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.047633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.047647 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.049040 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.069224 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.084511 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.099940 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.117679 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.133810 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.149882 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.150526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.150583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.150599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.150621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.150640 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.164530 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.188460 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.203570 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.217157 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.231006 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.244775 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.253311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.253348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.253370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.253387 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.253399 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.260822 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.281259 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc7daef4ce6549abbb4aca4c3638247df1b2b51f5a4c99bfecf194de13d66afa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"message\\\":\\\" for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:02Z is after 2025-08-24T17:21:41Z]\\\\nI1003 13:31:02.825480 6368 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1003 13:31:02.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:19Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.360042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.360106 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.360123 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.360147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.360164 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.462975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.463048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.463065 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.463090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.463108 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.566260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.566325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.566338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.566357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.566370 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.670610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.670698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.670724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.670753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.670772 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.684984 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.685061 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:19 crc kubenswrapper[4959]: E1003 13:31:19.685149 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:19 crc kubenswrapper[4959]: E1003 13:31:19.685365 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.774345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.774487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.774565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.774602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.774625 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.877866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.877921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.877935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.877953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.877966 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.980985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.981077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.981086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.981110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.981120 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:19Z","lastTransitionTime":"2025-10-03T13:31:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.994869 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/2.log" Oct 03 13:31:19 crc kubenswrapper[4959]: I1003 13:31:19.999279 4959 scope.go:117] "RemoveContainer" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" Oct 03 13:31:19 crc kubenswrapper[4959]: E1003 13:31:19.999699 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.013679 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.029258 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.046331 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.067604 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.081952 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.083950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.084049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.084066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.084089 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.084106 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.095249 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.113422 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.129280 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.142786 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.155527 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.167285 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.182017 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.187406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.187478 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.187496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.187532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.187556 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.196397 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.212529 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.237378 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.253895 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.270015 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:20Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.290492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.290553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.290566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.290585 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.290601 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.397024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.397113 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.397128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.397166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.397179 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.500403 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.500480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.500498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.500523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.500541 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.603989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.604047 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.604079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.604097 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.604109 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.684602 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.684619 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:20 crc kubenswrapper[4959]: E1003 13:31:20.684805 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:20 crc kubenswrapper[4959]: E1003 13:31:20.684946 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.706907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.706973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.706985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.707009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.707022 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.810019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.810083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.810103 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.810127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.810147 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.912432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.912491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.912506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.912526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:20 crc kubenswrapper[4959]: I1003 13:31:20.912544 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:20Z","lastTransitionTime":"2025-10-03T13:31:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.014066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.014105 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.014119 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.014134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.014143 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.023294 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:21 crc kubenswrapper[4959]: E1003 13:31:21.023473 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:21 crc kubenswrapper[4959]: E1003 13:31:21.023539 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:31:37.023521524 +0000 UTC m=+66.226864941 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.117158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.117215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.117224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.117238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.117248 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.220395 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.220430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.220438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.220451 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.220460 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.323130 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.323184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.323244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.323269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.323286 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.425491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.426043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.426139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.426246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.426321 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.529750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.529793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.529806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.529824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.529835 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.632428 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.632488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.632508 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.632534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.632553 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.685296 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.685435 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:21 crc kubenswrapper[4959]: E1003 13:31:21.686256 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:21 crc kubenswrapper[4959]: E1003 13:31:21.686454 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.703297 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.725127 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.736043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.736107 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.736119 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.736142 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.736156 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.739306 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.756786 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.782974 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.804214 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.826253 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.838720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.838771 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.838783 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.838803 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.838815 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.841596 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.858675 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.881000 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.897535 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.914003 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.925866 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.939808 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.941567 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.941654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.941666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.941692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.941706 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:21Z","lastTransitionTime":"2025-10-03T13:31:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.955362 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.969600 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:21 crc kubenswrapper[4959]: I1003 13:31:21.981233 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:21Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.043915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.043985 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.043998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.044016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.044029 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.146936 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.146977 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.146990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.147010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.147021 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.249975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.250023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.250032 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.250051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.250060 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.352152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.352208 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.352218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.352234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.352245 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.454953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.455001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.455012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.455030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.455043 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.538466 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.538800 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:31:54.538755039 +0000 UTC m=+83.742098456 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.558841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.558882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.558893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.558912 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.558921 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.639523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.639900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.639961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.639992 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640007 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640038 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640052 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640150 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640175 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640257 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640264 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:54.640142278 +0000 UTC m=+83.843485695 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640316 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:54.640302521 +0000 UTC m=+83.843645938 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640146 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640359 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:54.640351434 +0000 UTC m=+83.843694851 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640511 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.640548 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:31:54.640538178 +0000 UTC m=+83.843881595 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.661718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.661770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.661782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.661802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.661815 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.685147 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.685311 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.685446 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:22 crc kubenswrapper[4959]: E1003 13:31:22.685772 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.764062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.764352 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.764437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.764525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.764599 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.867148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.867223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.867233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.867252 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.867264 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.970373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.970460 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.970479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.970504 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:22 crc kubenswrapper[4959]: I1003 13:31:22.970522 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:22Z","lastTransitionTime":"2025-10-03T13:31:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.073718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.073765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.073775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.073794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.073804 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.180015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.180075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.180086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.180104 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.180125 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.282604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.282730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.282748 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.282774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.282793 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.385313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.385355 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.385366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.385390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.385403 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.488564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.488610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.488621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.488641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.488656 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.592625 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.592674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.592683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.592706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.592718 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.684796 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.684866 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:23 crc kubenswrapper[4959]: E1003 13:31:23.684984 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:23 crc kubenswrapper[4959]: E1003 13:31:23.685066 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.694865 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.694901 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.694913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.694932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.694950 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.798221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.798546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.798644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.798758 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.798858 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.902178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.902277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.902292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.902334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:23 crc kubenswrapper[4959]: I1003 13:31:23.902347 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:23Z","lastTransitionTime":"2025-10-03T13:31:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.005790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.005850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.005870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.005900 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.005927 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.109797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.109862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.109872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.109898 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.109910 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.212912 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.212997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.213023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.213053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.213071 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.316891 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.317336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.317632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.317821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.317998 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.420578 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.420637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.420653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.420679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.420696 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.523994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.524045 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.524057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.524079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.524102 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.626793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.626857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.626873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.626897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.626918 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.684826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.684826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:24 crc kubenswrapper[4959]: E1003 13:31:24.685036 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:24 crc kubenswrapper[4959]: E1003 13:31:24.685147 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.729610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.729680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.729692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.729709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.729721 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.833114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.833179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.833241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.833284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.833303 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.835145 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.848642 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.851674 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.868505 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.885652 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.901434 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.913711 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.931568 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.935761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.935800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.935813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.935833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.935847 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:24Z","lastTransitionTime":"2025-10-03T13:31:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.946106 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.966016 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.984138 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:24 crc kubenswrapper[4959]: I1003 13:31:24.997783 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:24Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.014666 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.027047 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.038970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.039035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.039051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.039072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.039087 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.043391 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.059351 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.074166 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.090828 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.112840 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:25Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.141747 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.141804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.141821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.141842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.141857 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.245524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.245591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.245608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.245633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.245648 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.348743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.348824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.348847 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.348879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.348903 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.451545 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.451597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.451622 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.451643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.451654 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.554257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.554313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.554321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.554337 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.554346 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.657109 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.657180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.657248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.657278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.657300 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.685185 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.685351 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:25 crc kubenswrapper[4959]: E1003 13:31:25.685486 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:25 crc kubenswrapper[4959]: E1003 13:31:25.685642 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.760442 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.760499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.760515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.760535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.760550 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.863870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.863933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.863950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.863974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.863989 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.966948 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.967018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.967043 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.967075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:25 crc kubenswrapper[4959]: I1003 13:31:25.967099 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:25Z","lastTransitionTime":"2025-10-03T13:31:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.069741 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.069807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.069823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.069848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.069865 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.173383 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.173467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.173494 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.173529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.173552 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.276342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.276402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.276412 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.276435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.276447 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.378593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.378644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.378657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.378678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.378692 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.481087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.481158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.481171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.481218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.481230 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.583587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.583624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.583633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.583647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.583656 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.684760 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:26 crc kubenswrapper[4959]: E1003 13:31:26.685276 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.684760 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:26 crc kubenswrapper[4959]: E1003 13:31:26.685776 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.686999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.687044 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.687054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.687073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.687089 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.790260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.790312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.790323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.790343 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.790358 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.893010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.893050 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.893058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.893076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.893087 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.995926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.995981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.995993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.996012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:26 crc kubenswrapper[4959]: I1003 13:31:26.997724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:26Z","lastTransitionTime":"2025-10-03T13:31:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.101140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.101218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.101233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.101283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.101299 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.205095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.205169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.205232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.205260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.205278 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.308277 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.308328 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.308341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.308360 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.308373 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.412435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.412494 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.412511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.412536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.412552 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.515141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.515290 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.515316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.515345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.515369 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.618371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.618424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.618433 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.618450 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.618462 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.684560 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:27 crc kubenswrapper[4959]: E1003 13:31:27.684734 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.684793 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:27 crc kubenswrapper[4959]: E1003 13:31:27.684979 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.721327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.721373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.721385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.721401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.721412 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.824430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.824496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.824509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.824531 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.824543 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.928765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.928846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.928874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.928913 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:27 crc kubenswrapper[4959]: I1003 13:31:27.928937 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:27Z","lastTransitionTime":"2025-10-03T13:31:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.032170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.032300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.032317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.032341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.032353 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.133989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.134067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.134079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.134099 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.134113 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.148603 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:28Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.152388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.152441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.152449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.152469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.152481 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.166248 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:28Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.171048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.171116 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.171142 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.171165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.171180 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.186169 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:28Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.190874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.190911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.190921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.190937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.190947 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.202945 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:28Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.207385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.207425 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.207436 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.207454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.207465 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.221795 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:28Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.221914 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.224063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.224114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.224128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.224148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.224162 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.326828 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.326874 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.326886 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.326907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.326920 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.429990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.430029 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.430038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.430055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.430065 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.532860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.532903 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.532922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.532942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.532951 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.636431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.637099 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.637285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.637386 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.637458 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.684870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.684925 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.685043 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:28 crc kubenswrapper[4959]: E1003 13:31:28.685125 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.739800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.740046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.740103 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.740171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.740248 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.843679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.843721 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.843731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.843750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.843780 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.946706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.946744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.946751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.946767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:28 crc kubenswrapper[4959]: I1003 13:31:28.946776 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:28Z","lastTransitionTime":"2025-10-03T13:31:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.049814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.049858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.049872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.049890 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.049900 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.153938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.154015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.154030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.154058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.154077 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.257576 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.257628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.257641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.257660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.257671 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.362020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.362550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.362669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.362784 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.362878 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.465645 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.465689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.465700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.465719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.465731 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.568593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.568629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.568638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.568656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.568665 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.670990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.671046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.671054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.671073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.671087 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.684903 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.684952 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:29 crc kubenswrapper[4959]: E1003 13:31:29.685081 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:29 crc kubenswrapper[4959]: E1003 13:31:29.685288 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.774309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.774356 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.774367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.774390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.774403 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.877921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.877981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.877996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.878023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.878040 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.981218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.981287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.981303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.981326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:29 crc kubenswrapper[4959]: I1003 13:31:29.981339 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:29Z","lastTransitionTime":"2025-10-03T13:31:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.084027 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.084093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.084109 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.084132 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.084149 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.187243 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.187302 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.187313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.187330 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.187341 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.289466 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.289525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.289539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.289559 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.289572 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.392539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.392587 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.392598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.392616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.392628 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.495826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.496582 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.496620 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.496643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.496655 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.599489 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.599572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.599583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.599600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.599611 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.684949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.684956 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:30 crc kubenswrapper[4959]: E1003 13:31:30.685234 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:30 crc kubenswrapper[4959]: E1003 13:31:30.685339 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.702154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.702225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.702236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.702257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.702268 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.804885 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.804932 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.804946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.804965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.804975 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.907909 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.907960 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.907976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.907994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:30 crc kubenswrapper[4959]: I1003 13:31:30.908006 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:30Z","lastTransitionTime":"2025-10-03T13:31:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.010566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.010603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.010612 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.010627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.010638 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.113546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.113594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.113605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.113621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.113632 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.216616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.216656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.216676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.216695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.216706 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.319215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.319255 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.319265 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.319281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.319293 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.422114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.422156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.422166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.422183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.422209 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.524361 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.524411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.524421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.524441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.524452 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.626697 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.626733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.626740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.626755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.626764 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.685046 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.685063 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:31 crc kubenswrapper[4959]: E1003 13:31:31.685297 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:31 crc kubenswrapper[4959]: E1003 13:31:31.685387 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.696639 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.711659 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.726003 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.729258 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.729309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.729318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.729332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.729341 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.737183 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.759254 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.772450 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.786988 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.799998 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.812684 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.830025 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.831148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.831213 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.831223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.831238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.831247 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.841982 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.852781 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.869323 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.889502 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.906099 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.915741 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.925150 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.933642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.933695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.933704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.933720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.933730 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:31Z","lastTransitionTime":"2025-10-03T13:31:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:31 crc kubenswrapper[4959]: I1003 13:31:31.943231 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:31Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.035147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.035186 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.035214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.035230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.035244 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.137672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.137923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.137990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.138064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.138125 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.240846 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.241230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.241243 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.241261 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.241270 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.343694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.343736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.343745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.343760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.343770 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.446384 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.446435 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.446446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.446465 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.446481 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.549174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.549458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.549527 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.549602 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.549669 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.652093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.652185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.652223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.652246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.652265 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.685008 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.685170 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:32 crc kubenswrapper[4959]: E1003 13:31:32.686794 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:32 crc kubenswrapper[4959]: E1003 13:31:32.686993 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.755218 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.755274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.755286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.755304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.755316 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.858134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.858214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.858226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.858243 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.858254 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.960754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.960794 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.960802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.960822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:32 crc kubenswrapper[4959]: I1003 13:31:32.960831 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:32Z","lastTransitionTime":"2025-10-03T13:31:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.063971 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.064016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.064025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.064040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.064050 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.167341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.167386 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.167396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.167413 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.167425 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.270024 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.270082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.270091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.270107 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.270115 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.376621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.376680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.376692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.376709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.376719 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.479829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.479940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.479953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.479976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.479993 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.582295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.582331 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.582339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.582352 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.582362 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.684946 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:33 crc kubenswrapper[4959]: E1003 13:31:33.685124 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.685171 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:33 crc kubenswrapper[4959]: E1003 13:31:33.685403 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.685910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.685999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.686061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.686125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.686184 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.789166 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.789273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.789285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.789305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.789318 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.891683 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.891734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.891744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.891760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.891769 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.994034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.994080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.994092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.994111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:33 crc kubenswrapper[4959]: I1003 13:31:33.994122 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:33Z","lastTransitionTime":"2025-10-03T13:31:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.096682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.096719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.096729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.096746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.096757 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.199136 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.199204 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.199213 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.199231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.199243 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.301552 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.301589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.301597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.301611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.301618 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.404546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.404586 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.404594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.404608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.404617 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.508096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.508439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.508583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.508745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.508902 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.611653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.611691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.611699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.611713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.611722 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.685223 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.685250 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:34 crc kubenswrapper[4959]: E1003 13:31:34.685652 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:34 crc kubenswrapper[4959]: E1003 13:31:34.685790 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.685921 4959 scope.go:117] "RemoveContainer" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" Oct 03 13:31:34 crc kubenswrapper[4959]: E1003 13:31:34.686082 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.714664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.714774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.714788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.714811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.714825 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.817323 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.817387 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.817407 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.817430 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.817448 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.921090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.921148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.921158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.921176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:34 crc kubenswrapper[4959]: I1003 13:31:34.921187 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:34Z","lastTransitionTime":"2025-10-03T13:31:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.023125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.023186 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.023225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.023246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.023262 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.125488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.125535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.125546 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.125563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.125576 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.228613 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.228658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.228670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.228687 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.228699 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.331711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.331801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.331833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.331869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.331896 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.434575 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.434631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.434646 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.434671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.434688 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.537994 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.538057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.538070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.538092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.538104 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.642469 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.642526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.642536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.642552 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.642561 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.684759 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.684783 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:35 crc kubenswrapper[4959]: E1003 13:31:35.684970 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:35 crc kubenswrapper[4959]: E1003 13:31:35.685049 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.745747 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.745800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.745809 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.745826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.745836 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.848424 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.848470 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.848482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.848498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.848508 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.951164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.951250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.951264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.951284 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:35 crc kubenswrapper[4959]: I1003 13:31:35.951297 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:35Z","lastTransitionTime":"2025-10-03T13:31:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.053699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.053745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.053754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.053773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.053784 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.156396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.156463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.156476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.156498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.156514 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.258898 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.258938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.258950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.258965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.258973 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.361271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.361310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.361319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.361334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.361343 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.463299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.463341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.463351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.463385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.463394 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.565400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.565448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.565463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.565482 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.565494 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.668038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.668088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.668099 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.668116 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.668126 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.684789 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.684856 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:36 crc kubenswrapper[4959]: E1003 13:31:36.684972 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:36 crc kubenswrapper[4959]: E1003 13:31:36.685112 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.770353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.770393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.770402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.770416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.770424 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.872817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.872854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.872864 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.872878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.872887 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.975894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.975973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.975988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.976010 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:36 crc kubenswrapper[4959]: I1003 13:31:36.976027 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:36Z","lastTransitionTime":"2025-10-03T13:31:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.078997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.079046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.079057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.079076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.079088 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.093515 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:37 crc kubenswrapper[4959]: E1003 13:31:37.093697 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:37 crc kubenswrapper[4959]: E1003 13:31:37.093759 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:32:09.093742577 +0000 UTC m=+98.297085994 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.181070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.181110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.181118 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.181134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.181143 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.283920 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.284213 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.284309 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.284397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.284484 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.387588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.387649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.387660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.387681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.387693 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.490291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.490332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.490345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.490364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.490375 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.593640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.593685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.593722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.593736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.593748 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.684992 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.685087 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:37 crc kubenswrapper[4959]: E1003 13:31:37.685166 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:37 crc kubenswrapper[4959]: E1003 13:31:37.685312 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.695774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.696026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.696108 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.696227 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.696321 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.799260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.799305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.799315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.799331 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.799341 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.903095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.903160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.903178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.903230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:37 crc kubenswrapper[4959]: I1003 13:31:37.903253 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:37Z","lastTransitionTime":"2025-10-03T13:31:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.006144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.006220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.006232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.006250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.006261 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.108515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.108562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.108573 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.108591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.108605 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.210719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.210754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.210762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.210776 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.210785 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.231663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.231698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.231705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.231718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.231726 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.245349 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:38Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.248590 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.248637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.248653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.248711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.248726 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.260664 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:38Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.264023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.264060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.264069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.264084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.264093 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.275109 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:38Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.278294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.278329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.278340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.278358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.278369 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.289121 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:38Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.294978 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.295028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.295042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.295064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.295077 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.307374 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:38Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.307498 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.313170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.313245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.313257 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.313295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.313307 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.415348 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.415625 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.415722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.415813 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.415901 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.518293 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.518512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.518605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.518710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.518803 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.620320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.620354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.620365 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.620382 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.620393 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.684380 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.684377 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.684706 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:38 crc kubenswrapper[4959]: E1003 13:31:38.684782 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.722307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.722346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.722358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.722374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.722383 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.824544 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.824592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.824603 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.824623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.824635 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.928672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.928719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.928729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.928750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:38 crc kubenswrapper[4959]: I1003 13:31:38.928762 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:38Z","lastTransitionTime":"2025-10-03T13:31:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.031169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.031267 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.031283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.031302 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.031315 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.133591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.133636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.133649 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.133666 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.133680 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.236256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.236305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.236317 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.236335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.236350 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.339040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.339110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.339122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.339139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.339150 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.441396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.441669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.441753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.441850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.441955 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.544253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.544332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.544346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.544366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.544399 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.651379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.651494 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.651516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.651543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.651563 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.685578 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.685672 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:39 crc kubenswrapper[4959]: E1003 13:31:39.685839 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:39 crc kubenswrapper[4959]: E1003 13:31:39.685985 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.753941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.753993 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.754006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.754030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.754046 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.856579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.856640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.856656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.856681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.856698 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.959126 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.959159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.959170 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.959214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:39 crc kubenswrapper[4959]: I1003 13:31:39.959227 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:39Z","lastTransitionTime":"2025-10-03T13:31:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060715 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/0.log" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060796 4959 generic.go:334] "Generic (PLEG): container finished" podID="117b68eb-14c1-456d-8c37-6259f7328587" containerID="68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8" exitCode=1 Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060853 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerDied","Data":"68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.060960 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.061483 4959 scope.go:117] "RemoveContainer" containerID="68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.074415 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.098607 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.109814 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.123987 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.136089 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.147429 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.161905 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.164106 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.164278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.164385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.164474 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.164556 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.173739 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.184724 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.198476 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.209069 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.222150 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.231710 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.241636 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.259773 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.267150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.267201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.267210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.267231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.267241 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.271054 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.282472 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.294637 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:40Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.370179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.370417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.370476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.370619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.370677 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.472574 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.472606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.472615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.472629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.472637 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.574589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.574659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.574671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.574690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.574701 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.676910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.677230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.677342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.677422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.677492 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.685505 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.685600 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:40 crc kubenswrapper[4959]: E1003 13:31:40.685686 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:40 crc kubenswrapper[4959]: E1003 13:31:40.685901 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.780026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.780060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.780068 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.780082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.780091 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.882600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.883351 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.883400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.883421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.883435 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.986107 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.986151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.986160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.986176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:40 crc kubenswrapper[4959]: I1003 13:31:40.986203 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:40Z","lastTransitionTime":"2025-10-03T13:31:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.065344 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/0.log" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.065684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerStarted","Data":"7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.080546 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.088468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.088539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.088554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.088577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.088593 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.094845 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.105214 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.118658 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.131619 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.145243 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.157057 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.167914 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.180242 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.191312 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.191377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.191393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.191417 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.191433 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.192407 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.204233 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.221515 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.233592 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.245850 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.257074 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.267912 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.279159 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.294307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.294364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.294378 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.294399 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.294416 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.300979 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.397621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.397672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.397684 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.397702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.397714 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.499818 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.499868 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.499880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.499899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.499913 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.602238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.602294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.602307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.602327 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.602339 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.685533 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:41 crc kubenswrapper[4959]: E1003 13:31:41.685701 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.685784 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:41 crc kubenswrapper[4959]: E1003 13:31:41.686069 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.701179 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.704939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.704975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.704988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.705006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.705021 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.713794 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.725624 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.736505 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.752797 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.765292 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.777993 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.789828 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.800691 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.807286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.807313 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.807322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.807335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.807344 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.813856 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.826184 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.837271 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.854015 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.862779 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.873508 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.886590 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.896919 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.909535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.909570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.909581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.909598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.909607 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:41Z","lastTransitionTime":"2025-10-03T13:31:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:41 crc kubenswrapper[4959]: I1003 13:31:41.912591 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:41Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.011716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.011757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.011768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.011785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.011797 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.114698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.114749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.114762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.114780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.114794 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.217904 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.218183 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.218301 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.218390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.218467 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.321376 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.321419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.321427 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.321442 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.321454 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.424127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.424179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.424429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.424453 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.424714 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.527120 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.527164 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.527184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.527216 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.527226 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.630180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.630241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.630251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.630266 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.630274 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.684949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.685073 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:42 crc kubenswrapper[4959]: E1003 13:31:42.685279 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:42 crc kubenswrapper[4959]: E1003 13:31:42.685083 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.696565 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.732792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.732830 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.732848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.732878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.732887 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.835528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.835800 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.835812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.835829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.835841 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.937560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.937612 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.937621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.937635 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:42 crc kubenswrapper[4959]: I1003 13:31:42.937643 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:42Z","lastTransitionTime":"2025-10-03T13:31:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.039570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.039610 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.039622 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.039639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.039650 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.141933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.141980 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.141991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.142009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.142021 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.244016 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.244056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.244066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.244080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.244091 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.346651 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.346697 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.346709 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.346726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.346738 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.448911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.448965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.448976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.448997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.449011 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.551233 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.551271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.551283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.551300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.551311 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.653629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.653670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.653682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.653699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.653710 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.685043 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:43 crc kubenswrapper[4959]: E1003 13:31:43.685168 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.685443 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:43 crc kubenswrapper[4959]: E1003 13:31:43.685605 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.755946 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.756006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.756015 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.756036 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.756046 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.858591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.858657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.858671 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.858695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.858711 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.961796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.961837 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.961848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.961866 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:43 crc kubenswrapper[4959]: I1003 13:31:43.961880 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:43Z","lastTransitionTime":"2025-10-03T13:31:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.064644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.064681 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.064690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.064705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.064713 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.167034 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.167073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.167086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.167103 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.167114 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.269147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.269211 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.269223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.269250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.269264 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.371593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.371631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.371639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.371657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.371666 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.473708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.473745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.473752 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.473766 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.473775 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.578408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.578757 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.578878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.578997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.579112 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.681924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.681968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.681979 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.681995 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.682010 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.685082 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.685142 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:44 crc kubenswrapper[4959]: E1003 13:31:44.685576 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:44 crc kubenswrapper[4959]: E1003 13:31:44.685764 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.784457 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.784912 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.784986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.785070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.785134 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.887788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.887833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.887843 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.887860 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.887871 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.990831 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.990876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.990884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.990901 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:44 crc kubenswrapper[4959]: I1003 13:31:44.990910 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:44Z","lastTransitionTime":"2025-10-03T13:31:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.092940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.092975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.092983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.092996 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.093004 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.195728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.195775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.195783 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.195797 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.195807 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.298019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.298064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.298075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.298092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.298105 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.400688 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.400738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.400750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.400772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.400784 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.504037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.504081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.504092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.504110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.504123 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.606768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.606814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.606825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.606842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.606852 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.684521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:45 crc kubenswrapper[4959]: E1003 13:31:45.684649 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.684521 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:45 crc kubenswrapper[4959]: E1003 13:31:45.684728 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.708957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.709033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.709054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.709080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.709096 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.811565 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.811614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.811628 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.811647 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.811658 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.914229 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.914278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.914291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.914308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:45 crc kubenswrapper[4959]: I1003 13:31:45.914322 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:45Z","lastTransitionTime":"2025-10-03T13:31:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.016952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.016999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.017012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.017031 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.017044 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.120217 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.120522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.120606 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.120877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.120948 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.224252 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.224344 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.224368 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.224401 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.224422 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.328119 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.328569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.328774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.328982 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.329131 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.432419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.432489 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.432512 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.432543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.432563 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.535702 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.535774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.535795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.535823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.535841 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.639227 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.639304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.639329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.639363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.639386 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.685454 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.685533 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:46 crc kubenswrapper[4959]: E1003 13:31:46.685674 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:46 crc kubenswrapper[4959]: E1003 13:31:46.685795 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.742870 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.742937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.742962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.742992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.743011 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.845028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.845065 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.845074 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.845093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.845101 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.949492 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.949624 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.949643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.949662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:46 crc kubenswrapper[4959]: I1003 13:31:46.949679 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:46Z","lastTransitionTime":"2025-10-03T13:31:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.052600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.053171 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.053320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.053412 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.053488 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.156012 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.156062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.156077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.156098 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.156112 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.258566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.258605 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.258615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.258630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.258639 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.361548 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.361795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.361931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.362062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.362179 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.465629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.465686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.465698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.465718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.465731 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.569493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.569863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.569957 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.570055 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.570173 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.673511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.673570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.673583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.673601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.673614 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.685180 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.685267 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:47 crc kubenswrapper[4959]: E1003 13:31:47.685365 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:47 crc kubenswrapper[4959]: E1003 13:31:47.685415 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.776382 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.776787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.776989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.777149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.777374 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.879722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.879773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.879789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.879808 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.879825 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.982515 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.982552 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.982562 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.982579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:47 crc kubenswrapper[4959]: I1003 13:31:47.982589 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:47Z","lastTransitionTime":"2025-10-03T13:31:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.085899 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.085973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.086002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.086039 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.086063 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.189047 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.189086 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.189095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.189109 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.189120 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.291825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.292072 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.292172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.292279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.292375 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.395707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.396031 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.396121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.396278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.396407 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.420554 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.420591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.420600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.420615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.420625 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.432241 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:48Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.436991 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.437048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.437062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.437085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.437098 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.450072 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:48Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.455765 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.455812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.455827 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.455850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.455864 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.471568 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:48Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.476623 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.476686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.476728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.476750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.476762 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.490913 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:48Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.500611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.500664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.500679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.500698 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.500708 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.519429 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:48Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.519925 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.522581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.523007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.523090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.523172 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.523337 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.626611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.626662 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.626673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.626691 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.626703 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.685057 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.685539 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.685624 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:48 crc kubenswrapper[4959]: E1003 13:31:48.685760 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.685786 4959 scope.go:117] "RemoveContainer" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.730149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.730234 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.730251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.730273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.730289 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.833231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.833274 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.833286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.833306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.833318 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.935841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.935895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.935905 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.935921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:48 crc kubenswrapper[4959]: I1003 13:31:48.935933 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:48Z","lastTransitionTime":"2025-10-03T13:31:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.038447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.038480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.038488 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.038505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.038517 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.093570 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/2.log" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.096756 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.097349 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.109634 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.126974 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.137818 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.140639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.140689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.140700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.140718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.140728 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.151696 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.164936 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.177925 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.190286 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.202408 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.216957 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.231510 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.242292 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.243276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.243326 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.243339 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.243358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.243371 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.253973 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.267528 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5abcd84-0b4e-4de4-928a-433804aa2103\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.290741 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.306424 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.319052 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.333651 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.346426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.346471 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.346483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.346502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.346514 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.348178 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.360798 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:49Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.449247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.449283 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.449294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.449311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.449322 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.551804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.551840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.551848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.551863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.551871 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.654630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.654670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.654679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.654695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.654705 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.684859 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.684929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:49 crc kubenswrapper[4959]: E1003 13:31:49.685077 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:49 crc kubenswrapper[4959]: E1003 13:31:49.685221 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.756677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.756722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.756730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.756743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.756751 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.859619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.859674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.859692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.859715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.859731 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.962295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.962357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.962380 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.962411 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:49 crc kubenswrapper[4959]: I1003 13:31:49.962435 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:49Z","lastTransitionTime":"2025-10-03T13:31:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.065838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.065877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.065888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.065907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.065918 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.103591 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/3.log" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.104366 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/2.log" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.107441 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" exitCode=1 Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.107493 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.107536 4959 scope.go:117] "RemoveContainer" containerID="1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.108185 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:31:50 crc kubenswrapper[4959]: E1003 13:31:50.108354 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.127831 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.143463 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.156375 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.167665 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.169740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.169805 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.169822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.169845 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.169860 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.185471 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1087ab41476757d0e1e0db29908c83035aa1b01bebae7cf98f515ffbddac1fb9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:18Z\\\",\\\"message\\\":\\\"go:208] Removed *v1.Pod event handler 6\\\\nI1003 13:31:18.729438 6588 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 13:31:18.729449 6588 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 13:31:18.729460 6588 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 13:31:18.729491 6588 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 13:31:18.729507 6588 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 13:31:18.729520 6588 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 13:31:18.733976 6588 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 13:31:18.734130 6588 factory.go:656] Stopping watch factory\\\\nI1003 13:31:18.734130 6588 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 13:31:18.734288 6588 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 13:31:18.734334 6588 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:49Z\\\",\\\"message\\\":\\\"go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1003 13:31:49.464184 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nI1003 13:31:49.464266 6952 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-cluster-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103074996 seconds. No OVN measurement.\\\\nI1003 13:31:49.464144 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1003 13:31:49.464285 6952 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nF1003 13:31:49.464290 6952 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"http\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.197824 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.209878 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.224782 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.239213 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.253007 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.268245 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.271810 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.271850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.271861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.271878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.271889 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.282896 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.298996 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.311571 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.324773 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.337408 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5abcd84-0b4e-4de4-928a-433804aa2103\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.355427 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.369934 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.373876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.373925 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.373937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.373956 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.373967 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.381030 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:50Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.476272 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.476355 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.476373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.476394 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.476409 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.579026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.579082 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.579093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.579111 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.579122 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.681712 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.681753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.681762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.681780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.681789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.684944 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.684965 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:50 crc kubenswrapper[4959]: E1003 13:31:50.685052 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:50 crc kubenswrapper[4959]: E1003 13:31:50.685174 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.784880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.784936 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.784947 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.784965 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.784982 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.888224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.888447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.888467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.888490 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.888501 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.991676 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.991720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.991738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.991754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:50 crc kubenswrapper[4959]: I1003 13:31:50.991766 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:50Z","lastTransitionTime":"2025-10-03T13:31:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.094256 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.094528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.094591 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.094653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.094717 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.112311 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/3.log" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.116262 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:31:51 crc kubenswrapper[4959]: E1003 13:31:51.116601 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.134004 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.153824 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.165450 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.178066 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.189526 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5abcd84-0b4e-4de4-928a-433804aa2103\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.197140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.197214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.197227 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.197247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.197259 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.210416 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.224813 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.240694 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.256317 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.270837 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.284716 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.300764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.300814 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.300824 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.300842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.300852 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.304243 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:49Z\\\",\\\"message\\\":\\\"go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1003 13:31:49.464184 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nI1003 13:31:49.464266 6952 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-cluster-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103074996 seconds. No OVN measurement.\\\\nI1003 13:31:49.464144 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1003 13:31:49.464285 6952 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nF1003 13:31:49.464290 6952 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"http\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.319838 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.330752 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.340979 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.354627 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.366572 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.383207 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.394890 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.403703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.403746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.403755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.403769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.403779 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.506273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.506315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.506325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.506342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.506353 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.609586 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.609648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.609660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.609679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.609690 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.685118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:51 crc kubenswrapper[4959]: E1003 13:31:51.685301 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.685351 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:51 crc kubenswrapper[4959]: E1003 13:31:51.685489 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.705443 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.712180 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.712273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.712296 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.712329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.712352 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.721439 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5abcd84-0b4e-4de4-928a-433804aa2103\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.745515 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.757953 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.769906 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.783978 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.793973 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.803820 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.815061 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.815334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.815445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.815537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.815638 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.816632 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.832818 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.844833 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.864477 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:49Z\\\",\\\"message\\\":\\\"go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1003 13:31:49.464184 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nI1003 13:31:49.464266 6952 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-cluster-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103074996 seconds. No OVN measurement.\\\\nI1003 13:31:49.464144 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1003 13:31:49.464285 6952 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nF1003 13:31:49.464290 6952 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"http\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.880137 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.892444 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.904450 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.916688 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.918211 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.918269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.918280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.918297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.918306 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:51Z","lastTransitionTime":"2025-10-03T13:31:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.930901 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.942758 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:51 crc kubenswrapper[4959]: I1003 13:31:51.952867 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:51Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.021057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.021362 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.021456 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.021564 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.021665 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.123737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.123788 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.123821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.123838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.123851 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.226002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.226303 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.226321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.226341 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.226354 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.329101 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.329149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.329158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.329174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.329183 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.431332 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.431381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.431392 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.431410 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.431420 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.534247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.534314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.534325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.534394 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.534407 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.636949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.636988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.637000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.637017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.637030 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.685021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.685058 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:52 crc kubenswrapper[4959]: E1003 13:31:52.685165 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:52 crc kubenswrapper[4959]: E1003 13:31:52.685310 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.739127 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.739176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.739205 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.739220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.739229 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.841319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.841358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.841366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.841381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.841390 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.943940 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.943988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.944000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.944018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:52 crc kubenswrapper[4959]: I1003 13:31:52.944032 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:52Z","lastTransitionTime":"2025-10-03T13:31:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.047179 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.047253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.047276 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.047292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.047305 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.149613 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.149675 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.149685 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.149700 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.149709 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.252165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.252230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.252239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.252253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.252263 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.355231 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.355278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.355289 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.355307 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.355318 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.457703 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.457751 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.457787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.457804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.457815 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.562826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.562892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.562910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.562937 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.562955 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.666182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.666259 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.666273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.666295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.666308 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.685467 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.685616 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:53 crc kubenswrapper[4959]: E1003 13:31:53.686238 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:53 crc kubenswrapper[4959]: E1003 13:31:53.686254 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.768633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.768915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.768999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.769123 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.769263 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.871353 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.871630 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.871706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.871775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.871836 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.974054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.974097 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.974108 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.974125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:53 crc kubenswrapper[4959]: I1003 13:31:53.974138 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:53Z","lastTransitionTime":"2025-10-03T13:31:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.076670 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.076939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.077144 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.077368 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.077565 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.180635 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.181239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.181358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.181454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.181544 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.283967 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.284020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.284033 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.284051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.284062 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.386093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.386145 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.386156 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.386174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.386203 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.488627 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.488673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.488714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.488731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.488741 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.576776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.576967 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.576928105 +0000 UTC m=+147.780271542 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.591840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.591879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.591888 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.591904 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.591915 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.677542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.677595 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.677619 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.677641 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677778 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677800 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677813 4959 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677837 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677889 4959 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677911 4959 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677922 4959 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.677861 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.677845298 +0000 UTC m=+147.881188715 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.678015 4959 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.678022 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.677996382 +0000 UTC m=+147.881339839 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.678120 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.678096154 +0000 UTC m=+147.881439611 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.678147 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.678130405 +0000 UTC m=+147.881473852 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.684479 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.684593 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.684657 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:54 crc kubenswrapper[4959]: E1003 13:31:54.684806 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.695160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.695222 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.695232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.695248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.695258 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.797736 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.797787 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.797817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.797835 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.797848 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.900829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.900892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.900908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.900927 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:54 crc kubenswrapper[4959]: I1003 13:31:54.900939 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:54Z","lastTransitionTime":"2025-10-03T13:31:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.005661 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.005779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.005804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.005833 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.005853 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.108153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.108215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.108224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.108239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.108249 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.212137 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.212216 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.212230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.212247 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.212257 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.315150 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.315236 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.315249 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.315273 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.315290 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.418840 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.418901 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.418917 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.418939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.418954 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.522102 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.522165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.522173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.522201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.522211 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.625026 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.625112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.625138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.625168 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.625268 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.684758 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:55 crc kubenswrapper[4959]: E1003 13:31:55.684907 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.684986 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:55 crc kubenswrapper[4959]: E1003 13:31:55.685276 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.728035 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.728090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.728112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.728138 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.728156 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.830887 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.830938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.830963 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.831020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.831034 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.934668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.934713 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.934724 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.934742 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:55 crc kubenswrapper[4959]: I1003 13:31:55.934754 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:55Z","lastTransitionTime":"2025-10-03T13:31:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.037810 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.037881 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.037897 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.037921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.037938 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.140597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.140656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.140673 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.140695 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.140711 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.243760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.243801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.243812 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.243827 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.243837 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.349570 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.349640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.349654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.349677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.349693 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.453345 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.453381 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.453389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.453404 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.453412 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.555357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.555400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.555408 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.555426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.555435 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.658373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.658415 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.658425 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.658441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.658454 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.685173 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.685281 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:56 crc kubenswrapper[4959]: E1003 13:31:56.685387 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:56 crc kubenswrapper[4959]: E1003 13:31:56.685537 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.761014 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.761062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.761073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.761090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.761102 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.864017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.864079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.864100 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.864133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.864155 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.966447 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.966506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.966529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.966558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:56 crc kubenswrapper[4959]: I1003 13:31:56.966582 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:56Z","lastTransitionTime":"2025-10-03T13:31:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.069013 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.069056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.069069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.069087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.069100 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.172069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.172115 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.172129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.172147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.172159 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.275669 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.275714 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.275725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.275743 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.275754 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.378873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.378934 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.378954 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.378979 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.378997 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.481882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.482487 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.482644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.482792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.482965 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.586250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.586318 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.586334 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.586364 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.586380 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.684935 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:57 crc kubenswrapper[4959]: E1003 13:31:57.685136 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.685553 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:57 crc kubenswrapper[4959]: E1003 13:31:57.685671 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.688732 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.688766 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.688777 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.688792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.688804 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.791069 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.791107 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.791118 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.791133 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.791144 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.893263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.893300 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.893308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.893322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.893331 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.996306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.996406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.996438 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.996472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:57 crc kubenswrapper[4959]: I1003 13:31:57.996493 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:57Z","lastTransitionTime":"2025-10-03T13:31:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.100419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.100480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.100495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.100513 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.100526 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.203734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.203782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.203790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.203806 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.203815 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.306825 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.306879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.306894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.306915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.306934 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.409930 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.409978 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.409989 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.410006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.410018 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.513858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.513922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.513941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.513978 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.514005 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.616371 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.616431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.616449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.616472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.616486 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.685481 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.685571 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.685938 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.686110 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.719485 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.719577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.719621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.719668 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.719694 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.795975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.796037 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.796052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.796080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.796100 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.814416 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.820046 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.820104 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.820121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.820139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.820151 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.838914 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.842707 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.842745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.842756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.842770 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.842781 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.860926 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.865479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.865577 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.865601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.865633 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.865659 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.882050 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.887448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.887501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.887514 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.887537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.887554 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.902000 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:31:58Z is after 2025-08-24T17:21:41Z" Oct 03 13:31:58 crc kubenswrapper[4959]: E1003 13:31:58.902158 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.904279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.904315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.904325 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.904340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:58 crc kubenswrapper[4959]: I1003 13:31:58.904350 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:58Z","lastTransitionTime":"2025-10-03T13:31:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.008169 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.008295 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.008315 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.008338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.008358 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.111319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.111363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.111379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.111402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.111421 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.214454 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.214513 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.214532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.214563 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.214579 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.323223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.323311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.323322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.323347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.323357 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.426260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.426299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.426308 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.426320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.426332 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.528569 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.528617 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.528626 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.528641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.528652 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.631738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.631808 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.631819 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.631841 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.631855 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.684575 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.684701 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:31:59 crc kubenswrapper[4959]: E1003 13:31:59.684740 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:31:59 crc kubenswrapper[4959]: E1003 13:31:59.684928 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.734853 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.734942 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.734961 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.734982 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.735004 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.837958 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.838007 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.838022 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.838044 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.838060 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.940791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.940838 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.940849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.940867 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:31:59 crc kubenswrapper[4959]: I1003 13:31:59.940879 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:31:59Z","lastTransitionTime":"2025-10-03T13:31:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.043821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.043877 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.043889 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.043908 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.043922 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.146690 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.146750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.146766 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.146785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.146802 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.249528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.249592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.249614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.249644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.249666 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.352678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.352820 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.352842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.352872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.352894 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.456097 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.456140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.456151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.456167 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.456178 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.558795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.558850 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.558863 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.558880 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.558892 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.660923 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.660976 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.661020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.661040 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.661049 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.685445 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.685493 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:00 crc kubenswrapper[4959]: E1003 13:32:00.685622 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:00 crc kubenswrapper[4959]: E1003 13:32:00.685731 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.764319 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.764361 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.764369 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.764385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.764397 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.866938 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.866988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.866999 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.867018 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.867028 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.969585 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.969644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.969657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.969678 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:00 crc kubenswrapper[4959]: I1003 13:32:00.969689 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:00Z","lastTransitionTime":"2025-10-03T13:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.072875 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.072933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.072949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.072972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.072986 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.176260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.176306 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.176320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.176342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.176357 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.278738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.278780 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.278791 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.278807 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.278818 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.381154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.381293 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.381311 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.381342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.381361 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.484042 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.484088 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.484096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.484114 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.484124 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.586655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.586716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.586733 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.586759 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.586777 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.685223 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.685264 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:01 crc kubenswrapper[4959]: E1003 13:32:01.685352 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:01 crc kubenswrapper[4959]: E1003 13:32:01.685439 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.689654 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.689718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.689732 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.689754 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.689768 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.703597 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.719767 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89ec17a3b953cc858cd27593aee8cd1f7868c446045a79510eb1b23a878eeb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65f49991c8ae319fa57c11822a8a163316204863d542e581585af5a2e7aaec27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.735998 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5cd61a029ca3fc78643f2a33f8faf9f7fa80b6442168792f1dcd51a0bd2bde4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-plgmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4dsjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.753068 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zskn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"416afb5d-c0cb-4285-bd76-d7419c63cd54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ab1f5f6500e59966ab6c132f156213914ae4850345933997e779d9812da1dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2285t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zskn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.769435 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v679m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc425287-c256-4549-80ed-62c460cbe88e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30a3d3d3f74d2ba8a32a3324ce09cab6c7d01558623ba300fceedddf6e75c27c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69624bd41f4e40c00c283998c7c171383bc739b3a1cdf8c96c2ab638fc504122\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae21f69ee092289d3bff55a215873ddfa945ea41c4172913b8ff65c0a6b0e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8aaf36ec6b3982e00556c18baf9fd03e0bf8d4e7fd8e1aea4507f72988eaa4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4c137a3bb34e6fee1faa9f603aa6e1e5f8cba668f77c2fa05eab7e3ae132adb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59c51e39c9a3cfbec50b85634d0a4dab3fc8fb675f6dadd3fbd3a7324970db4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45e58311f023440d72813f82028d811d4f3916f9692f3fb01e4fc26343238d7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4knxj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v679m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.791927 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f5b506-7c54-4068-a367-b4da4c61ad03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7200d478d09d0a1c04e96ede3053f5b030cd8dc1abff0177d23ab85f532b379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4703b83a8a48e5c0d8ec62fc9dc8877f8214ac5d7350427f191565fa5192f37d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://433dc7b3234c40d2a62bfc8438f30bbcf5f24c27f44a6486f2aff0857dffb64f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27573f45d5f7d91c36c3be4e6696b139978f46385b060b92fe654e38b8fcdf5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e4e4f76b97c9132bc6e1dd89dc2e44241342de4d8e72b3139baca47c40b2844\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T13:30:45Z\\\",\\\"message\\\":\\\"W1003 13:30:34.766818 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 13:30:34.767858 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759498234 cert, and key in /tmp/serving-cert-4058549150/serving-signer.crt, /tmp/serving-cert-4058549150/serving-signer.key\\\\nI1003 13:30:35.039265 1 observer_polling.go:159] Starting file observer\\\\nW1003 13:30:35.046270 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 13:30:35.050143 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 13:30:35.054645 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4058549150/tls.crt::/tmp/serving-cert-4058549150/tls.key\\\\\\\"\\\\nF1003 13:30:45.378226 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cc3392af5d0b881677303ccb7af54a368e8a1e3bf2f5d811283feb4a0f3a2fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e25914e4d69845f1b6accbb3933c2cbbdd452445c253a98c8a74dbf91f5840\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.792694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.792738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.792746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.792768 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.792779 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.810950 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.830795 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f5c88fcce4223227a5844fafa558bfbfb9b612273429b208e035760d77d8bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.848884 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c76e698387152183764dd4986c0cef34312bf732434871abd2e49df9ee2fb514\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.868794 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-tfz2x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"117b68eb-14c1-456d-8c37-6259f7328587\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:39Z\\\",\\\"message\\\":\\\"2025-10-03T13:30:54+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb\\\\n2025-10-03T13:30:54+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_52236800-d61c-4233-a8f0-0df14d7d8ccb to /host/opt/cni/bin/\\\\n2025-10-03T13:30:54Z [verbose] multus-daemon started\\\\n2025-10-03T13:30:54Z [verbose] Readiness Indicator file check\\\\n2025-10-03T13:31:39Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wcdz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-tfz2x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.880973 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9gj84" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22c9cebf-ffbf-4281-9200-58029e483f64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bbc92ae4fc077f20cdaddb2505f01ffb7acb22a725b88a039e54882c8f0dbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pczm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9gj84\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.893231 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da45dabb-f9ff-428b-904c-f45a1cf701a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26a6b64f198028cdc883ccec2f8da116746feb2b776411ed8b0fc324c6c366b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09547a83b1670727b9cbfcad67a972001111541f8539ba9d6d123c3c902df6c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:31:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rnvk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mxxqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.895445 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.895496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.895509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.895530 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.895543 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.907328 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5abcd84-0b4e-4de4-928a-433804aa2103\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5150e939e904fd97c174cc3020e196978e281d6e2f5d890bf2c3487dda24f8f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c5f9e76773dd9571e7efd3b96c1377873427edb99f1654fe31535aa86c152c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.930148 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a60b1e44-2539-4c2b-a498-6ba3541dcc85\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe47e2f5a5d49958b877cb51bbdb57abf0186926e9c505afeaa239dc78dd9d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f0baca5144fe6aec09d746c892aebe233524a05b248b429104e69c6566f4416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e911e37eb5c5e2cf5025588149af4c6ad78f75049c2e72263a8c8cddab58cf6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a02c4e770119ab7cbc680b5f11413d70e0708ec912e914b12910da7b345679a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1da31561fa3ed822f336801df4b6f2be3b37a0f257be0fbe362503540538a965\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1af17ebcb1914ab01896bb8b32585006440965d5990b841b8db2a2456e5651da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b7de3595a0900e59109ca07cd0eb6503c9729bee08d8cab85f6dd1af84b2ffd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ae61b8e451b8d737b820ba597418f8d0b285e8b8392641be51d1442982e3db8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.943749 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3a679d1-7412-4843-998e-8d1c109bf7ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k4shq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:31:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gf4tr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.960013 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f5e6e38-bc00-4299-a88c-69c4ac2b73e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a6e62ade9b3536fb2c145bdf200b9c3f312329bb4aeaccf45baa7d45ce9ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87a9f7e974eedb4abd61961c4c9013cd2c36e5006888289df1d92b566f6f277a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a274885518a602c6d839afc5a1106d3189d2dd623c14499ce09a9f56b161bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13849d46d43f2b57f7092dbe79fd3d281fa707d42a74cecf3e8bcd151450ecd6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.972731 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.984381 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf0b4c6-badb-44c1-9472-74c5fd52bd1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:31:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e8c8cefea56fb8e156881958c6cf65e4ec381f36a143403d7eda2b9c87c62d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://20b54ab0adf3fac952955ca2dda5602fe57b030a07c92a578c42bc50310c3760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://608b30cdce59a53d137bedfb819ea8c6dbe6d8c8c829d7d279b25484935902ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9ea040045d610b6557285adefabe50e6c9ee096dd49511a2f72d1c1c56d4765\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:01Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.999028 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.999090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.999106 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.999129 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:01 crc kubenswrapper[4959]: I1003 13:32:01.999142 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:01Z","lastTransitionTime":"2025-10-03T13:32:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.008056 4959 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3185a69a-b160-4257-afe1-a2165aa6e961\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T13:31:49Z\\\",\\\"message\\\":\\\"go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1003 13:31:49.464184 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nI1003 13:31:49.464266 6952 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-cluster-machine-approver/machine-approver. OVN-Kubernetes controller took 0.103074996 seconds. No OVN measurement.\\\\nI1003 13:31:49.464144 6952 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1003 13:31:49.464285 6952 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-xqbk4\\\\nF1003 13:31:49.464290 6952 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"http\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T13:31:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T13:30:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T13:30:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T13:30:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tjvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T13:30:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xqbk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:02Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.102558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.103001 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.103084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.103160 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.103256 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.206621 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.207052 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.207132 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.207279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.207392 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.309921 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.310477 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.310595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.310725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.310812 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.414354 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.414394 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.414405 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.414418 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.414429 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.516667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.516726 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.516738 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.516755 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.516770 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.619362 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.619420 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.619439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.619462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.619477 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.684654 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.684693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:02 crc kubenswrapper[4959]: E1003 13:32:02.684826 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:02 crc kubenswrapper[4959]: E1003 13:32:02.684920 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.721895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.721941 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.721953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.721969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.721982 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.824095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.824128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.824136 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.824149 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.824158 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.927002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.927070 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.927085 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.927104 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:02 crc kubenswrapper[4959]: I1003 13:32:02.927116 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:02Z","lastTransitionTime":"2025-10-03T13:32:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.030093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.030146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.030159 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.030177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.030217 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.132682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.132722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.132731 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.132744 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.132752 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.235719 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.235761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.235774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.235795 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.235812 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.338429 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.338475 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.338484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.338502 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.338514 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.441541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.441588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.441599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.441615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.441626 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.544112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.544151 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.544161 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.544178 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.544205 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.646811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.646857 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.646868 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.646884 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.646894 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.684782 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.684865 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:03 crc kubenswrapper[4959]: E1003 13:32:03.684920 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:03 crc kubenswrapper[4959]: E1003 13:32:03.685043 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.749643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.749689 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.749699 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.749715 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.749724 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.852079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.852402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.852472 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.852550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.852626 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.955608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.955964 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.956112 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.956288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:03 crc kubenswrapper[4959]: I1003 13:32:03.956437 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:03Z","lastTransitionTime":"2025-10-03T13:32:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.058842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.058883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.058895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.058915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.058927 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.161458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.161510 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.161522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.161541 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.161553 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.264467 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.264511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.264523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.264542 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.264555 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.367363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.367805 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.367886 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.368019 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.368092 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.470526 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.470581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.470594 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.470611 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.470623 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.573868 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.573935 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.573949 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.573973 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.573989 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.677536 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.677588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.677599 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.677619 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.677630 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.685353 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.685364 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:04 crc kubenswrapper[4959]: E1003 13:32:04.685486 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:04 crc kubenswrapper[4959]: E1003 13:32:04.686128 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.687291 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:32:04 crc kubenswrapper[4959]: E1003 13:32:04.687588 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.779706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.779753 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.779764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.779782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.779792 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.882223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.882260 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.882269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.882282 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.882291 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.985448 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.985491 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.985501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.985518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:04 crc kubenswrapper[4959]: I1003 13:32:04.985529 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:04Z","lastTransitionTime":"2025-10-03T13:32:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.089782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.089852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.089869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.089894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.089911 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.192664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.192745 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.192769 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.192801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.192824 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.295839 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.295892 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.295907 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.295924 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.295937 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.398716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.398763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.398774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.398790 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.398801 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.501238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.501316 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.501344 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.501370 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.501387 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.605030 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.605081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.605093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.605110 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.605121 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.684860 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.685065 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:05 crc kubenswrapper[4959]: E1003 13:32:05.685305 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:05 crc kubenswrapper[4959]: E1003 13:32:05.685421 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.707386 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.707431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.707441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.707461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.707474 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.810148 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.810201 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.810210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.810228 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.810242 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.913000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.913076 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.913098 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.913132 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:05 crc kubenswrapper[4959]: I1003 13:32:05.913154 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:05Z","lastTransitionTime":"2025-10-03T13:32:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.015215 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.015254 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.015265 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.015281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.015293 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.117796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.117849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.117873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.117894 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.117908 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.220808 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.220852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.220862 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.220878 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.220888 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.323167 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.323230 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.323241 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.323259 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.323271 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.425581 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.425629 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.425639 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.425655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.425664 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.527792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.527844 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.527856 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.527876 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.527890 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.630225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.630496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.630507 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.630522 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.630534 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.685486 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:06 crc kubenswrapper[4959]: E1003 13:32:06.685676 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.685482 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:06 crc kubenswrapper[4959]: E1003 13:32:06.685894 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.733717 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.733772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.733782 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.733803 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.733815 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.836125 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.836185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.836225 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.836246 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.836259 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.938781 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.938842 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.938852 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.938872 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:06 crc kubenswrapper[4959]: I1003 13:32:06.938883 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:06Z","lastTransitionTime":"2025-10-03T13:32:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.041746 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.041792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.041804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.041822 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.041835 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.145226 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.145294 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.145305 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.145322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.145337 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.247939 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.247990 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.248002 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.248021 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.248035 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.351419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.351496 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.351534 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.351557 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.351570 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.454521 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.454583 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.454592 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.454608 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.454618 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.560181 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.560251 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.560266 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.560291 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.560311 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.663373 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.663422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.663431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.663449 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.663458 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.684753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.684798 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:07 crc kubenswrapper[4959]: E1003 13:32:07.684920 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:07 crc kubenswrapper[4959]: E1003 13:32:07.685016 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.766134 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.766173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.766184 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.766229 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.766240 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.870391 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.870452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.870463 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.870479 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.870493 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.972667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.972706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.972722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.972739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:07 crc kubenswrapper[4959]: I1003 13:32:07.972749 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:07Z","lastTransitionTime":"2025-10-03T13:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.075640 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.075706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.075728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.075756 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.075776 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.180173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.180271 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.180287 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.180314 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.180336 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.282970 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.283050 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.283060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.283077 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.283089 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.386093 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.386128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.386139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.386155 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.386166 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.488740 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.489121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.489393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.489597 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.489780 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.593558 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.593616 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.593634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.593659 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.593677 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.685141 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.685149 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:08 crc kubenswrapper[4959]: E1003 13:32:08.685796 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:08 crc kubenswrapper[4959]: E1003 13:32:08.685879 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.697269 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.697340 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.697358 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.697386 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.697403 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.799653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.799710 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.799723 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.799739 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.799748 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.902006 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.902051 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.902062 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.902081 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:08 crc kubenswrapper[4959]: I1003 13:32:08.902118 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:08Z","lastTransitionTime":"2025-10-03T13:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.004931 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.004972 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.004981 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.004998 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.005009 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.107476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.107518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.107528 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.107543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.107552 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.113043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.113186 4959 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.113260 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs podName:e3a679d1-7412-4843-998e-8d1c109bf7ec nodeName:}" failed. No retries permitted until 2025-10-03 13:33:13.113247496 +0000 UTC m=+162.316590913 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs") pod "network-metrics-daemon-gf4tr" (UID: "e3a679d1-7412-4843-998e-8d1c109bf7ec") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.210458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.210520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.210537 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.210560 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.210576 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.221005 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.221054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.221066 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.221083 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.221095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.233979 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:09Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.237950 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.237992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.238003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.238023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.238034 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.254384 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:09Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.258992 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.259044 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.259060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.259084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.259096 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.273055 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:09Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.277221 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.277253 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.277262 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.277278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.277287 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.290046 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:09Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.295152 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.295210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.295223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.295244 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.295255 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.311176 4959 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T13:32:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9ec5ca6-9088-4c22-bd9a-b3c75487e558\\\",\\\"systemUUID\\\":\\\"168d26af-4a20-4818-904a-4d599a720a2a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T13:32:09Z is after 2025-08-24T17:21:41Z" Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.311373 4959 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.313525 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.313595 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.313612 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.313634 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.313651 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.420802 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.421462 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.421476 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.421498 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.421512 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.523658 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.523728 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.523750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.523779 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.523802 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.626934 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.627004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.627023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.627048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.627066 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.685659 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.685816 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.685992 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:09 crc kubenswrapper[4959]: E1003 13:32:09.686134 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.730232 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.730297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.730322 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.730346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.730364 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.833636 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.833706 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.833730 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.833761 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.833780 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.936484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.936538 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.936551 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.936571 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:09 crc kubenswrapper[4959]: I1003 13:32:09.936584 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:09Z","lastTransitionTime":"2025-10-03T13:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.039330 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.039367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.039374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.039389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.039398 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.142336 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.142379 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.142389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.142406 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.142415 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.245004 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.245049 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.245067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.245087 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.245099 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.348446 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.348493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.348505 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.348523 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.348535 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.451383 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.451432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.451444 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.451461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.451472 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.554331 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.554367 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.554375 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.554389 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.554399 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.657094 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.657146 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.657158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.657177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.657211 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.685473 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.685534 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:10 crc kubenswrapper[4959]: E1003 13:32:10.685727 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:10 crc kubenswrapper[4959]: E1003 13:32:10.685909 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.760817 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.760879 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.760895 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.760918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.760931 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.864586 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.864632 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.864642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.864660 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.864671 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.967655 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.967720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.967737 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.967760 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:10 crc kubenswrapper[4959]: I1003 13:32:10.967776 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:10Z","lastTransitionTime":"2025-10-03T13:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.070593 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.070653 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.070667 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.070686 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.070698 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.173186 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.173250 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.173268 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.173288 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.173298 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.276355 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.276416 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.276432 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.276461 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.276478 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.379279 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.379346 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.379363 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.379388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.379405 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.481849 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.481906 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.481922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.481944 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.481960 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.585173 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.585286 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.585338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.585362 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.585383 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.684797 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.684797 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:11 crc kubenswrapper[4959]: E1003 13:32:11.684981 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:11 crc kubenswrapper[4959]: E1003 13:32:11.685131 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.694139 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.694796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.695025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.695280 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.695489 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.732655 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.732626803 podStartE2EDuration="1m22.732626803s" podCreationTimestamp="2025-10-03 13:30:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.73083823 +0000 UTC m=+100.934181657" watchObservedRunningTime="2025-10-03 13:32:11.732626803 +0000 UTC m=+100.935970270" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.794374 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podStartSLOduration=80.794330506 podStartE2EDuration="1m20.794330506s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.793275461 +0000 UTC m=+100.996618888" watchObservedRunningTime="2025-10-03 13:32:11.794330506 +0000 UTC m=+100.997673973" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.798067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.798113 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.798122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.798140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.798150 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.823210 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zskn9" podStartSLOduration=81.823180495 podStartE2EDuration="1m21.823180495s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.823177985 +0000 UTC m=+101.026521442" watchObservedRunningTime="2025-10-03 13:32:11.823180495 +0000 UTC m=+101.026523912" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.886357 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-v679m" podStartSLOduration=80.886336424 podStartE2EDuration="1m20.886336424s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.874167329 +0000 UTC m=+101.077510746" watchObservedRunningTime="2025-10-03 13:32:11.886336424 +0000 UTC m=+101.089679841" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.900750 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.900832 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.900843 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.900861 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.900872 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:11Z","lastTransitionTime":"2025-10-03T13:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.919558 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=29.919536787 podStartE2EDuration="29.919536787s" podCreationTimestamp="2025-10-03 13:31:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.886730473 +0000 UTC m=+101.090073890" watchObservedRunningTime="2025-10-03 13:32:11.919536787 +0000 UTC m=+101.122880204" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.919775 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=77.919768574 podStartE2EDuration="1m17.919768574s" podCreationTimestamp="2025-10-03 13:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.919412384 +0000 UTC m=+101.122755811" watchObservedRunningTime="2025-10-03 13:32:11.919768574 +0000 UTC m=+101.123111991" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.967546 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-tfz2x" podStartSLOduration=80.96752533 podStartE2EDuration="1m20.96752533s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.967265803 +0000 UTC m=+101.170609220" watchObservedRunningTime="2025-10-03 13:32:11.96752533 +0000 UTC m=+101.170868747" Oct 03 13:32:11 crc kubenswrapper[4959]: I1003 13:32:11.977895 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-9gj84" podStartSLOduration=81.9778775 podStartE2EDuration="1m21.9778775s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.977403669 +0000 UTC m=+101.180747106" watchObservedRunningTime="2025-10-03 13:32:11.9778775 +0000 UTC m=+101.181220917" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.001589 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mxxqp" podStartSLOduration=81.001569754 podStartE2EDuration="1m21.001569754s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:11.987695328 +0000 UTC m=+101.191038745" watchObservedRunningTime="2025-10-03 13:32:12.001569754 +0000 UTC m=+101.204913171" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.003023 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.003067 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.003080 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.003096 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.003109 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.015853 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.015836819 podStartE2EDuration="1m18.015836819s" podCreationTimestamp="2025-10-03 13:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:12.004033434 +0000 UTC m=+101.207376851" watchObservedRunningTime="2025-10-03 13:32:12.015836819 +0000 UTC m=+101.219180236" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.037838 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.037820272 podStartE2EDuration="48.037820272s" podCreationTimestamp="2025-10-03 13:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:12.037390281 +0000 UTC m=+101.240733698" watchObservedRunningTime="2025-10-03 13:32:12.037820272 +0000 UTC m=+101.241163699" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.105566 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.105631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.105644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.105665 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.105678 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.207598 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.207638 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.207648 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.207663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.207671 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.310374 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.310422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.310437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.310458 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.310471 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.413553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.413622 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.413637 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.413656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.413706 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.516121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.516185 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.516220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.516239 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.516251 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.619396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.619431 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.619441 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.619455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.619464 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.684952 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.684977 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:12 crc kubenswrapper[4959]: E1003 13:32:12.685119 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:12 crc kubenswrapper[4959]: E1003 13:32:12.685271 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.721672 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.721767 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.721796 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.721821 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.721840 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.825121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.825210 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.825224 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.825242 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.825255 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.928380 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.928426 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.928437 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.928455 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:12 crc kubenswrapper[4959]: I1003 13:32:12.928468 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:12Z","lastTransitionTime":"2025-10-03T13:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.031053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.031095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.031106 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.031124 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.031137 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.133997 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.134038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.134048 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.134063 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.134075 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.236926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.236974 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.236983 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.237000 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.237010 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.339722 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.339792 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.339848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.339883 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.339906 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.442366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.442409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.442419 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.442434 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.442445 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.545409 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.545480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.545499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.545524 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.545541 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.648588 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.648644 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.648656 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.648674 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.648686 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.684915 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.684920 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:13 crc kubenswrapper[4959]: E1003 13:32:13.685241 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:13 crc kubenswrapper[4959]: E1003 13:32:13.685356 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.751677 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.751764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.751781 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.751837 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.751856 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.854694 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.854763 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.854784 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.854816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.854836 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.958223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.958292 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.958310 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.958335 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:13 crc kubenswrapper[4959]: I1003 13:32:13.958353 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:13Z","lastTransitionTime":"2025-10-03T13:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.062158 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.062285 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.062297 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.062321 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.062336 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.164329 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.164388 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.164397 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.164414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.164428 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.267057 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.267145 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.267162 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.267187 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.267260 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.370091 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.370140 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.370154 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.370174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.370216 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.472772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.472816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.472829 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.472854 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.472866 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.575716 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.575764 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.575774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.575789 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.575798 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.678858 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.678900 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.678911 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.678926 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.678936 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.685415 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.685429 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:14 crc kubenswrapper[4959]: E1003 13:32:14.685506 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:14 crc kubenswrapper[4959]: E1003 13:32:14.685603 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.781600 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.781663 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.781680 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.781704 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.781721 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.884387 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.884484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.884520 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.884550 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.884573 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.986725 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.986772 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.986785 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.986801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:14 crc kubenswrapper[4959]: I1003 13:32:14.986811 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:14Z","lastTransitionTime":"2025-10-03T13:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.089873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.089952 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.089975 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.090003 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.090028 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.192053 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.192092 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.192103 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.192121 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.192131 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.294729 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.294773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.294783 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.294801 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.294812 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.397465 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.397509 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.397518 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.397535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.397545 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.500553 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.500604 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.500615 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.500631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.500642 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.604105 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.604162 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.604177 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.604219 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.604239 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.684586 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.684644 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:15 crc kubenswrapper[4959]: E1003 13:32:15.684791 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:15 crc kubenswrapper[4959]: E1003 13:32:15.684909 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.707017 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.707090 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.707107 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.707128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.707139 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.810393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.810459 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.810480 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.810499 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.810513 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.913642 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.913692 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.913705 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.913734 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:15 crc kubenswrapper[4959]: I1003 13:32:15.913746 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:15Z","lastTransitionTime":"2025-10-03T13:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.016953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.017020 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.017038 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.017064 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.017095 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.120747 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.121302 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.121516 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.121723 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.121915 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.225422 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.225493 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.225511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.225535 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.225554 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.328165 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.328214 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.328223 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.328238 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.328247 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.431320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.431368 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.431377 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.431393 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.431404 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.534009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.534047 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.534054 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.534071 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.534085 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.636009 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.636075 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.636095 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.636132 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.636153 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.684844 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.684845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:16 crc kubenswrapper[4959]: E1003 13:32:16.684988 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:16 crc kubenswrapper[4959]: E1003 13:32:16.685051 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.738278 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.738347 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.738366 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.738390 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.738406 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.840873 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.840910 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.840918 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.840933 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.840942 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.944264 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.944320 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.944338 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.944357 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:16 crc kubenswrapper[4959]: I1003 13:32:16.944370 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:16Z","lastTransitionTime":"2025-10-03T13:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.048128 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.048174 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.048220 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.048245 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.048261 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.150823 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.150882 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.150893 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.150914 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.150927 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.254402 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.254815 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.254962 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.255141 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.255404 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.358532 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.358607 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.358631 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.358664 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.358686 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.462539 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.462589 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.462601 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.462620 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.462634 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.564643 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.565058 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.565147 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.565281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.565363 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.668342 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.668385 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.668396 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.668414 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.668425 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.684928 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.685379 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:17 crc kubenswrapper[4959]: E1003 13:32:17.685492 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:17 crc kubenswrapper[4959]: E1003 13:32:17.685720 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.771025 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.771079 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.771097 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.771122 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.771140 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.875031 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.875073 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.875084 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.875098 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.875106 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.977614 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.977853 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.977969 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.978060 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:17 crc kubenswrapper[4959]: I1003 13:32:17.978146 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:17Z","lastTransitionTime":"2025-10-03T13:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.080153 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.080421 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.080483 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.080549 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.080618 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.183304 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.183579 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.183708 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.183804 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.183892 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.286986 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.287299 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.287400 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.287495 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.287601 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.390679 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.390718 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.390727 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.390741 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.390751 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.493711 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.493749 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.493759 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.493774 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.493789 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.595848 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.595915 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.595928 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.595953 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.595986 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.685489 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.685513 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:18 crc kubenswrapper[4959]: E1003 13:32:18.685734 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:18 crc kubenswrapper[4959]: E1003 13:32:18.685868 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.699732 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.699811 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.699836 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.699869 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.699895 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.803452 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.803501 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.803511 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.803529 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.803540 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.906816 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.906885 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.906896 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.906922 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:18 crc kubenswrapper[4959]: I1003 13:32:18.906935 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:18Z","lastTransitionTime":"2025-10-03T13:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.010572 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.010641 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.010657 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.010682 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.010710 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.113580 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.114056 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.114124 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.114281 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.114410 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.217410 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.217468 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.217484 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.217506 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.217520 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.320543 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.320826 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.320896 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.320968 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.321033 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.423720 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.423762 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.423773 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.423793 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.423805 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.522439 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.522831 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.522988 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.523176 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.523407 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.540182 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.540240 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.540248 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.540263 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.540271 4959 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T13:32:19Z","lastTransitionTime":"2025-10-03T13:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.585361 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh"] Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.585819 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.588423 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.588450 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.588744 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.591571 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.628413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e95f4b-cd2c-475c-a8ec-cee71975ae78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.628611 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.628655 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01e95f4b-cd2c-475c-a8ec-cee71975ae78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.628684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.628788 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e95f4b-cd2c-475c-a8ec-cee71975ae78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.685429 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.686000 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:19 crc kubenswrapper[4959]: E1003 13:32:19.686145 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:19 crc kubenswrapper[4959]: E1003 13:32:19.686336 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.686812 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:32:19 crc kubenswrapper[4959]: E1003 13:32:19.687142 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xqbk4_openshift-ovn-kubernetes(3185a69a-b160-4257-afe1-a2165aa6e961)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730362 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730430 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01e95f4b-cd2c-475c-a8ec-cee71975ae78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730528 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730540 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e95f4b-cd2c-475c-a8ec-cee71975ae78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730627 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e95f4b-cd2c-475c-a8ec-cee71975ae78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.730989 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/01e95f4b-cd2c-475c-a8ec-cee71975ae78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.731680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/01e95f4b-cd2c-475c-a8ec-cee71975ae78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.736733 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01e95f4b-cd2c-475c-a8ec-cee71975ae78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.755384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01e95f4b-cd2c-475c-a8ec-cee71975ae78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ww8fh\" (UID: \"01e95f4b-cd2c-475c-a8ec-cee71975ae78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: I1003 13:32:19.902067 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" Oct 03 13:32:19 crc kubenswrapper[4959]: W1003 13:32:19.919884 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01e95f4b_cd2c_475c_a8ec_cee71975ae78.slice/crio-3bf295c8ac535b0d90ec72aeacf44fedbaf888ba5964fac72daa017b92e69828 WatchSource:0}: Error finding container 3bf295c8ac535b0d90ec72aeacf44fedbaf888ba5964fac72daa017b92e69828: Status 404 returned error can't find the container with id 3bf295c8ac535b0d90ec72aeacf44fedbaf888ba5964fac72daa017b92e69828 Oct 03 13:32:20 crc kubenswrapper[4959]: I1003 13:32:20.207151 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" event={"ID":"01e95f4b-cd2c-475c-a8ec-cee71975ae78","Type":"ContainerStarted","Data":"f44c4700a7f34b0dd28ed04496132f18acd3d3f4dfb26fcc0444111fc665038d"} Oct 03 13:32:20 crc kubenswrapper[4959]: I1003 13:32:20.207699 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" event={"ID":"01e95f4b-cd2c-475c-a8ec-cee71975ae78","Type":"ContainerStarted","Data":"3bf295c8ac535b0d90ec72aeacf44fedbaf888ba5964fac72daa017b92e69828"} Oct 03 13:32:20 crc kubenswrapper[4959]: I1003 13:32:20.225931 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ww8fh" podStartSLOduration=90.225909319 podStartE2EDuration="1m30.225909319s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:20.225326165 +0000 UTC m=+109.428669582" watchObservedRunningTime="2025-10-03 13:32:20.225909319 +0000 UTC m=+109.429252736" Oct 03 13:32:20 crc kubenswrapper[4959]: I1003 13:32:20.685402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:20 crc kubenswrapper[4959]: I1003 13:32:20.685409 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:20 crc kubenswrapper[4959]: E1003 13:32:20.685618 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:20 crc kubenswrapper[4959]: E1003 13:32:20.685728 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:21 crc kubenswrapper[4959]: I1003 13:32:21.684948 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:21 crc kubenswrapper[4959]: I1003 13:32:21.685056 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:21 crc kubenswrapper[4959]: E1003 13:32:21.686486 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:21 crc kubenswrapper[4959]: E1003 13:32:21.686733 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:22 crc kubenswrapper[4959]: I1003 13:32:22.685303 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:22 crc kubenswrapper[4959]: I1003 13:32:22.685353 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:22 crc kubenswrapper[4959]: E1003 13:32:22.685491 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:22 crc kubenswrapper[4959]: E1003 13:32:22.685636 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:23 crc kubenswrapper[4959]: I1003 13:32:23.684830 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:23 crc kubenswrapper[4959]: I1003 13:32:23.684830 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:23 crc kubenswrapper[4959]: E1003 13:32:23.685042 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:23 crc kubenswrapper[4959]: E1003 13:32:23.685149 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:24 crc kubenswrapper[4959]: I1003 13:32:24.684720 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:24 crc kubenswrapper[4959]: I1003 13:32:24.684793 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:24 crc kubenswrapper[4959]: E1003 13:32:24.684983 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:24 crc kubenswrapper[4959]: E1003 13:32:24.685069 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:25 crc kubenswrapper[4959]: I1003 13:32:25.684775 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:25 crc kubenswrapper[4959]: I1003 13:32:25.684920 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:25 crc kubenswrapper[4959]: E1003 13:32:25.685053 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:25 crc kubenswrapper[4959]: E1003 13:32:25.685138 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.229409 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/1.log" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.230063 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/0.log" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.230163 4959 generic.go:334] "Generic (PLEG): container finished" podID="117b68eb-14c1-456d-8c37-6259f7328587" containerID="7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0" exitCode=1 Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.230233 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerDied","Data":"7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0"} Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.230307 4959 scope.go:117] "RemoveContainer" containerID="68cf54dc10d1922d9ddb11438e14bb4512de9df90acbc458ebf66ba493875ce8" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.230940 4959 scope.go:117] "RemoveContainer" containerID="7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0" Oct 03 13:32:26 crc kubenswrapper[4959]: E1003 13:32:26.231246 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-tfz2x_openshift-multus(117b68eb-14c1-456d-8c37-6259f7328587)\"" pod="openshift-multus/multus-tfz2x" podUID="117b68eb-14c1-456d-8c37-6259f7328587" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.684574 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:26 crc kubenswrapper[4959]: E1003 13:32:26.684741 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:26 crc kubenswrapper[4959]: I1003 13:32:26.684582 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:26 crc kubenswrapper[4959]: E1003 13:32:26.684946 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:27 crc kubenswrapper[4959]: I1003 13:32:27.235436 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/1.log" Oct 03 13:32:27 crc kubenswrapper[4959]: I1003 13:32:27.684508 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:27 crc kubenswrapper[4959]: I1003 13:32:27.684632 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:27 crc kubenswrapper[4959]: E1003 13:32:27.684816 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:27 crc kubenswrapper[4959]: E1003 13:32:27.685043 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:28 crc kubenswrapper[4959]: I1003 13:32:28.685275 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:28 crc kubenswrapper[4959]: E1003 13:32:28.685400 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:28 crc kubenswrapper[4959]: I1003 13:32:28.686382 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:28 crc kubenswrapper[4959]: E1003 13:32:28.686619 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:29 crc kubenswrapper[4959]: I1003 13:32:29.684559 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:29 crc kubenswrapper[4959]: E1003 13:32:29.684704 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:29 crc kubenswrapper[4959]: I1003 13:32:29.684728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:29 crc kubenswrapper[4959]: E1003 13:32:29.684811 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:30 crc kubenswrapper[4959]: I1003 13:32:30.684997 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:30 crc kubenswrapper[4959]: I1003 13:32:30.685032 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:30 crc kubenswrapper[4959]: E1003 13:32:30.685249 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:30 crc kubenswrapper[4959]: E1003 13:32:30.685453 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:31 crc kubenswrapper[4959]: I1003 13:32:31.684963 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:31 crc kubenswrapper[4959]: I1003 13:32:31.685792 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:31 crc kubenswrapper[4959]: E1003 13:32:31.686012 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:31 crc kubenswrapper[4959]: E1003 13:32:31.686120 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:31 crc kubenswrapper[4959]: E1003 13:32:31.692136 4959 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 03 13:32:31 crc kubenswrapper[4959]: E1003 13:32:31.827393 4959 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 13:32:32 crc kubenswrapper[4959]: I1003 13:32:32.684693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:32 crc kubenswrapper[4959]: E1003 13:32:32.684937 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:32 crc kubenswrapper[4959]: I1003 13:32:32.684982 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:32 crc kubenswrapper[4959]: E1003 13:32:32.685031 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:32 crc kubenswrapper[4959]: I1003 13:32:32.685991 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.256626 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/3.log" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.259152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerStarted","Data":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.259491 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.283276 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podStartSLOduration=102.283255636 podStartE2EDuration="1m42.283255636s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:33.282350684 +0000 UTC m=+122.485694121" watchObservedRunningTime="2025-10-03 13:32:33.283255636 +0000 UTC m=+122.486599053" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.689419 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:33 crc kubenswrapper[4959]: E1003 13:32:33.689556 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.689763 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:33 crc kubenswrapper[4959]: E1003 13:32:33.689886 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.703858 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gf4tr"] Oct 03 13:32:33 crc kubenswrapper[4959]: I1003 13:32:33.703977 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:33 crc kubenswrapper[4959]: E1003 13:32:33.704064 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:34 crc kubenswrapper[4959]: I1003 13:32:34.684659 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:34 crc kubenswrapper[4959]: E1003 13:32:34.684797 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:35 crc kubenswrapper[4959]: I1003 13:32:35.685282 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:35 crc kubenswrapper[4959]: I1003 13:32:35.685312 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:35 crc kubenswrapper[4959]: E1003 13:32:35.686029 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:35 crc kubenswrapper[4959]: I1003 13:32:35.685364 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:35 crc kubenswrapper[4959]: E1003 13:32:35.686115 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:35 crc kubenswrapper[4959]: E1003 13:32:35.686249 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:36 crc kubenswrapper[4959]: I1003 13:32:36.685086 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:36 crc kubenswrapper[4959]: E1003 13:32:36.685267 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:36 crc kubenswrapper[4959]: E1003 13:32:36.828337 4959 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 13:32:37 crc kubenswrapper[4959]: I1003 13:32:37.685224 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:37 crc kubenswrapper[4959]: I1003 13:32:37.685324 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:37 crc kubenswrapper[4959]: E1003 13:32:37.685369 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:37 crc kubenswrapper[4959]: I1003 13:32:37.685222 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:37 crc kubenswrapper[4959]: E1003 13:32:37.685556 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:37 crc kubenswrapper[4959]: E1003 13:32:37.685680 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:38 crc kubenswrapper[4959]: I1003 13:32:38.684659 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:38 crc kubenswrapper[4959]: E1003 13:32:38.684903 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:39 crc kubenswrapper[4959]: I1003 13:32:39.685488 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:39 crc kubenswrapper[4959]: E1003 13:32:39.686080 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:39 crc kubenswrapper[4959]: I1003 13:32:39.686325 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:39 crc kubenswrapper[4959]: I1003 13:32:39.686553 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:39 crc kubenswrapper[4959]: E1003 13:32:39.687078 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:39 crc kubenswrapper[4959]: E1003 13:32:39.687513 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:39 crc kubenswrapper[4959]: I1003 13:32:39.687921 4959 scope.go:117] "RemoveContainer" containerID="7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0" Oct 03 13:32:40 crc kubenswrapper[4959]: I1003 13:32:40.283970 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/1.log" Oct 03 13:32:40 crc kubenswrapper[4959]: I1003 13:32:40.284306 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerStarted","Data":"d2f163aff2d8dfb8ab5462c2397a3b695da1e7a41334a2cb0dd2b7bb93530ea4"} Oct 03 13:32:40 crc kubenswrapper[4959]: I1003 13:32:40.684738 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:40 crc kubenswrapper[4959]: E1003 13:32:40.684927 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 13:32:41 crc kubenswrapper[4959]: I1003 13:32:41.684696 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:41 crc kubenswrapper[4959]: E1003 13:32:41.687107 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 13:32:41 crc kubenswrapper[4959]: I1003 13:32:41.687963 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:41 crc kubenswrapper[4959]: I1003 13:32:41.688110 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:41 crc kubenswrapper[4959]: E1003 13:32:41.688155 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gf4tr" podUID="e3a679d1-7412-4843-998e-8d1c109bf7ec" Oct 03 13:32:41 crc kubenswrapper[4959]: E1003 13:32:41.688417 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 13:32:42 crc kubenswrapper[4959]: I1003 13:32:42.684986 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:42 crc kubenswrapper[4959]: I1003 13:32:42.688670 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 03 13:32:42 crc kubenswrapper[4959]: I1003 13:32:42.689824 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.685112 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.685172 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.685265 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.688571 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.690723 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.690853 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 03 13:32:43 crc kubenswrapper[4959]: I1003 13:32:43.692098 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.347775 4959 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.406007 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n9cfz"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.407439 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.407855 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h7s26"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.409020 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.409876 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.419706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.421389 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.421909 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.426597 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.426616 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.426596 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.427087 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.427150 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.427370 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.427940 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.428427 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.430333 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.430776 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.433045 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s7kpd"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.433676 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.433826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.434484 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.436551 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.440616 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.440709 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.440907 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.441253 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.441721 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.447015 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.447097 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.447470 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.447574 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.448294 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.453993 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.454433 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.468127 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.468356 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.470025 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.475343 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.483597 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.483949 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.484134 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.511470 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.511956 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.519687 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.519763 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.519868 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.519877 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.533906 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.536180 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.540340 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.547334 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.547639 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z6rtc"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.548463 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.548751 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.564839 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.568240 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.569775 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ttm4r"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.570261 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.570366 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.570610 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.576703 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577157 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577459 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577467 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577653 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577706 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.577838 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578127 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578150 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578635 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578752 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578890 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.578697 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.579049 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.579175 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.579546 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.579958 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.580536 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581040 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581174 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581317 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581536 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581577 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581720 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581787 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581833 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.581732 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582155 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582242 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-audit\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w7pt\" (UniqueName: \"kubernetes.io/projected/fcf06d86-50e8-4978-81ad-d9a2351afc39-kube-api-access-5w7pt\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582530 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582526 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582579 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pwqb\" (UniqueName: \"kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582607 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-dir\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582639 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.583875 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.583966 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.584452 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.584536 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.585364 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.585423 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.582678 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.586661 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf06d86-50e8-4978-81ad-d9a2351afc39-serving-cert\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.586732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-service-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.586775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-node-pullsecrets\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.586805 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-config\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.592425 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.592466 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n9cfz"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.592483 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fzwkw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.593027 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.597263 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.597667 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.586836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.600123 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.600284 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.600320 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-images\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.600368 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.600415 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-encryption-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.601230 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.601541 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.601583 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpxrr\" (UniqueName: \"kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.601611 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l67h9\" (UniqueName: \"kubernetes.io/projected/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-kube-api-access-l67h9\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-policies\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602456 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602484 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602507 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602535 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602564 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602819 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.602909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603021 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-audit-dir\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603085 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603108 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5cdb1c45-cb05-4d7d-9d75-693e90227102-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603132 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-etcd-client\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603153 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-serving-cert\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-encryption-config\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603236 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603266 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-etcd-serving-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603351 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggp4d\" (UniqueName: \"kubernetes.io/projected/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-kube-api-access-ggp4d\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.603387 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4fg\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-kube-api-access-7w4fg\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604371 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-image-import-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604508 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-serving-cert\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604545 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604597 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604628 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604735 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604818 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604883 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6k7d\" (UniqueName: \"kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604923 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.604978 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-client\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftqb4\" (UniqueName: \"kubernetes.io/projected/5cdb1c45-cb05-4d7d-9d75-693e90227102-kube-api-access-ftqb4\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605092 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c24kn\" (UniqueName: \"kubernetes.io/projected/677b5491-6c28-4625-854b-881e25662a57-kube-api-access-c24kn\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605121 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605142 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6258\" (UniqueName: \"kubernetes.io/projected/5c097a19-754b-465c-8c78-fc8b9ea45526-kube-api-access-r6258\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-config\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605225 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.605260 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.613713 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.618250 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.618883 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.619352 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.620084 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.622244 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.622430 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.622467 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.622643 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.623040 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.624451 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.624566 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.625241 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.625452 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626027 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626206 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626379 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626453 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626530 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626613 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626757 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626785 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626962 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626987 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.627061 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.627153 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.627265 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.626057 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.629568 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.631510 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zsxvg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.634914 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.643392 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.643480 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.643879 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.644114 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.644609 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.644782 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.645147 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.645261 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.645620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.645988 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.647605 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.648215 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.652315 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.652968 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.653203 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.653336 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.653447 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.653514 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.653563 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.665491 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.669643 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.670278 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.672487 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.672749 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.674359 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.677815 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.681337 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.681620 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-k89zn"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.682497 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.682877 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.683889 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.684425 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v6d5j"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.685559 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.686423 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.687021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.687416 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.688651 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h7s26"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.690669 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.691009 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.691636 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.692427 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.693016 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.694359 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.694981 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.696019 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.696419 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.697454 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s7kpd"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.698716 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.699266 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.700005 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.700609 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.701559 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z6rtc"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.702701 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.703845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.703943 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2prt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.704754 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706060 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706317 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggp4d\" (UniqueName: \"kubernetes.io/projected/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-kube-api-access-ggp4d\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706421 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-etcd-serving-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706511 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4fg\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-kube-api-access-7w4fg\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-service-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706369 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706785 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.706892 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zcp5\" (UniqueName: \"kubernetes.io/projected/847fd56c-ec05-4627-bd43-49add0bdc748-kube-api-access-6zcp5\") pod \"downloads-7954f5f757-ttm4r\" (UID: \"847fd56c-ec05-4627-bd43-49add0bdc748\") " pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-image-import-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707217 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-serving-cert\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707518 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707718 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-serving-cert\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707833 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-client\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707936 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-config\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708145 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl2p6\" (UniqueName: \"kubernetes.io/projected/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-kube-api-access-nl2p6\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711346 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711398 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6k7d\" (UniqueName: \"kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/daf6334f-8900-455a-95d7-8cc149fecc58-machine-approver-tls\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711443 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6mc\" (UniqueName: \"kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711465 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-config\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708900 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-auth-proxy-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.710115 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711537 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.711574 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707569 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-etcd-serving-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.707981 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.708100 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-image-import-ca\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.709945 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.710170 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.710376 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-client\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712252 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xv6l\" (UniqueName: \"kubernetes.io/projected/13e04245-e686-4b2b-9b90-68659825e4ad-kube-api-access-9xv6l\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712279 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-serving-cert\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712344 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712374 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftqb4\" (UniqueName: \"kubernetes.io/projected/5cdb1c45-cb05-4d7d-9d75-693e90227102-kube-api-access-ftqb4\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712396 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c24kn\" (UniqueName: \"kubernetes.io/projected/677b5491-6c28-4625-854b-881e25662a57-kube-api-access-c24kn\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96gx2\" (UniqueName: \"kubernetes.io/projected/f7190415-01a1-416c-b7a8-7945d9f20796-kube-api-access-96gx2\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712451 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712469 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.712427 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713114 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713228 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713294 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713328 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6258\" (UniqueName: \"kubernetes.io/projected/5c097a19-754b-465c-8c78-fc8b9ea45526-kube-api-access-r6258\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713380 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ttm4r"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713443 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-config\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713649 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713823 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.713978 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714108 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714138 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714185 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-audit\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714274 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w7pt\" (UniqueName: \"kubernetes.io/projected/fcf06d86-50e8-4978-81ad-d9a2351afc39-kube-api-access-5w7pt\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714302 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714394 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714412 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7190415-01a1-416c-b7a8-7945d9f20796-serving-cert\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714462 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-config\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714815 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-serving-cert\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.714912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715022 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pwqb\" (UniqueName: \"kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-dir\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715110 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715076 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715136 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-dir\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715166 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715335 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf06d86-50e8-4978-81ad-d9a2351afc39-serving-cert\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715720 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715791 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-service-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.715990 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.716751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-service-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.716826 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-node-pullsecrets\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.716879 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-node-pullsecrets\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.716903 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-config\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.716942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717085 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-images\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717165 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717211 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717230 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717257 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717292 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-encryption-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpxrr\" (UniqueName: \"kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717352 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l67h9\" (UniqueName: \"kubernetes.io/projected/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-kube-api-access-l67h9\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717371 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-policies\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717391 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717411 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717433 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717450 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717470 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717490 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717510 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-trusted-ca\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-audit-dir\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717575 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f7190415-01a1-416c-b7a8-7945d9f20796-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-etcd-client\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717621 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-serving-cert\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717645 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717669 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5cdb1c45-cb05-4d7d-9d75-693e90227102-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717689 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6q67\" (UniqueName: \"kubernetes.io/projected/daf6334f-8900-455a-95d7-8cc149fecc58-kube-api-access-w6q67\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.717721 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-encryption-config\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.718002 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.718363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/677b5491-6c28-4625-854b-881e25662a57-audit\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.719125 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.719211 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/677b5491-6c28-4625-854b-881e25662a57-audit-dir\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.719347 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-config\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.719853 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-etcd-client\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.720161 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.720703 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.721060 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.721283 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.721919 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-images\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.721964 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-encryption-config\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.722384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.722794 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.723122 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.723129 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.723717 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5c097a19-754b-465c-8c78-fc8b9ea45526-audit-policies\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.724618 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf06d86-50e8-4978-81ad-d9a2351afc39-serving-cert\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.724939 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5cdb1c45-cb05-4d7d-9d75-693e90227102-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.725473 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf06d86-50e8-4978-81ad-d9a2351afc39-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.725590 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c097a19-754b-465c-8c78-fc8b9ea45526-serving-cert\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.726371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.727666 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.728409 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bg4hd"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.728417 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.729335 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.729328 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-etcd-client\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.730026 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.730300 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.731292 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/677b5491-6c28-4625-854b-881e25662a57-encryption-config\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.734380 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fzwkw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.735525 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zsxvg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.737639 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.737948 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.740545 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.746672 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.747061 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.750896 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.751724 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.752919 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.754614 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-v8x2x"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.756013 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-452zl"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.756249 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.756661 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.757866 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.759481 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.761106 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.762430 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.763846 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.765477 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2prt"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.766928 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.768487 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.769675 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.770326 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.771405 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.772437 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.774051 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-v8x2x"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.775137 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.776171 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.777266 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.778311 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.779573 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-452zl"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.780648 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.781738 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.782928 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v6d5j"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.783967 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-mw5qw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.784965 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.785535 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mw5qw"] Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.790804 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.810413 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.818414 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gndmp\" (UniqueName: \"kubernetes.io/projected/a398dbb8-67e9-4390-9a50-9c5dd2c62303-kube-api-access-gndmp\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.818602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.818710 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrm78\" (UniqueName: \"kubernetes.io/projected/3b5aee17-20ec-4adb-ad67-a239878063a6-kube-api-access-nrm78\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.818845 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-plugins-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.818946 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-config\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-auth-proxy-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819107 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e75817c-456c-4d5c-981d-c434adca9411-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819205 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj5sv\" (UniqueName: \"kubernetes.io/projected/3f3cf4c7-106d-4f21-9596-6bd2356af306-kube-api-access-gj5sv\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-config\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819405 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819509 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw8hl\" (UniqueName: \"kubernetes.io/projected/b8f35bb3-a509-49f1-a945-0a8170569c1d-kube-api-access-qw8hl\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819595 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8f35bb3-a509-49f1-a945-0a8170569c1d-proxy-tls\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819766 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/289092c0-4c05-4c6d-bb0f-d36cb003b23e-metrics-tls\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk5lh\" (UniqueName: \"kubernetes.io/projected/2cebb480-f94a-4fef-b128-281428436da2-kube-api-access-qk5lh\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819810 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0061d337-ce1d-4925-b9d4-c6805aafe7ec-metrics-tls\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e98663-cc71-4af8-8a92-446851e11efe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.819939 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngq6q\" (UniqueName: \"kubernetes.io/projected/0061d337-ce1d-4925-b9d4-c6805aafe7ec-kube-api-access-ngq6q\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820118 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9tvg\" (UniqueName: \"kubernetes.io/projected/bac8b263-5a8c-4621-a046-496c25f91286-kube-api-access-m9tvg\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820231 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838dd7d5-7935-4238-b000-4adb94d759c6-config\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-stats-auth\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820293 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820305 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-auth-proxy-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsxrm\" (UniqueName: \"kubernetes.io/projected/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-kube-api-access-fsxrm\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820384 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-socket-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820454 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-metrics-certs\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820492 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-config\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820593 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-profile-collector-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx28x\" (UniqueName: \"kubernetes.io/projected/b04c7d45-8cef-4c51-8821-0eac0a6f6b07-kube-api-access-lx28x\") pod \"migrator-59844c95c7-j5tgf\" (UID: \"b04c7d45-8cef-4c51-8821-0eac0a6f6b07\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-service-ca-bundle\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.820953 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6q67\" (UniqueName: \"kubernetes.io/projected/daf6334f-8900-455a-95d7-8cc149fecc58-kube-api-access-w6q67\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821127 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5aee17-20ec-4adb-ad67-a239878063a6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821170 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821206 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzlxj\" (UniqueName: \"kubernetes.io/projected/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-kube-api-access-wzlxj\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821232 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a398dbb8-67e9-4390-9a50-9c5dd2c62303-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821274 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwfx9\" (UniqueName: \"kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821304 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821335 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t945\" (UniqueName: \"kubernetes.io/projected/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-kube-api-access-2t945\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821379 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-serving-cert\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-client\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821420 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-config\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl2p6\" (UniqueName: \"kubernetes.io/projected/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-kube-api-access-nl2p6\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821461 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6mc\" (UniqueName: \"kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821485 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-default-certificate\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/daf6334f-8900-455a-95d7-8cc149fecc58-machine-approver-tls\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821521 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-mountpoint-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xv6l\" (UniqueName: \"kubernetes.io/projected/13e04245-e686-4b2b-9b90-68659825e4ad-kube-api-access-9xv6l\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821560 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-serving-cert\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821576 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821593 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96gx2\" (UniqueName: \"kubernetes.io/projected/f7190415-01a1-416c-b7a8-7945d9f20796-kube-api-access-96gx2\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821611 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bv9\" (UniqueName: \"kubernetes.io/projected/60628569-ded6-49cb-8bb3-b814b5afc84a-kube-api-access-h2bv9\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821636 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821658 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kdwj\" (UniqueName: \"kubernetes.io/projected/d24f1f11-7f2c-4917-8648-cf1cc03d324b-kube-api-access-2kdwj\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821682 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-srv-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821728 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821752 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821777 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821801 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-csi-data-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821837 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821874 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821923 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/df758465-1b0a-4ecb-b70c-c256357e4ec0-tmpfs\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821950 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrdhw\" (UniqueName: \"kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.821978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7190415-01a1-416c-b7a8-7945d9f20796-serving-cert\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822007 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-srv-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822055 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/838dd7d5-7935-4238-b000-4adb94d759c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822088 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db9df3df-527a-4573-b1ac-4bf2bd94444e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822114 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e98663-cc71-4af8-8a92-446851e11efe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822140 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grhwh\" (UniqueName: \"kubernetes.io/projected/67e98663-cc71-4af8-8a92-446851e11efe-kube-api-access-grhwh\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822163 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/289092c0-4c05-4c6d-bb0f-d36cb003b23e-trusted-ca\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822221 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cebb480-f94a-4fef-b128-281428436da2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822248 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkzrk\" (UniqueName: \"kubernetes.io/projected/db9df3df-527a-4573-b1ac-4bf2bd94444e-kube-api-access-rkzrk\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822270 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822293 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac8b263-5a8c-4621-a046-496c25f91286-config\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822317 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822343 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b5aee17-20ec-4adb-ad67-a239878063a6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822368 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-webhook-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-registration-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-trusted-ca\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822441 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e75817c-456c-4d5c-981d-c434adca9411-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822466 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgjtz\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-kube-api-access-dgjtz\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac8b263-5a8c-4621-a046-496c25f91286-serving-cert\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f7190415-01a1-416c-b7a8-7945d9f20796-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822547 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-images\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822572 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-proxy-tls\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822629 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822655 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822683 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e75817c-456c-4d5c-981d-c434adca9411-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822743 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822770 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838dd7d5-7935-4238-b000-4adb94d759c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-service-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822829 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2lmn\" (UniqueName: \"kubernetes.io/projected/704c868e-4e03-4945-9d88-382b59d8ffd1-kube-api-access-s2lmn\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822886 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zcp5\" (UniqueName: \"kubernetes.io/projected/847fd56c-ec05-4627-bd43-49add0bdc748-kube-api-access-6zcp5\") pod \"downloads-7954f5f757-ttm4r\" (UID: \"847fd56c-ec05-4627-bd43-49add0bdc748\") " pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-965z4\" (UniqueName: \"kubernetes.io/projected/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-kube-api-access-965z4\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.822961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq6kj\" (UniqueName: \"kubernetes.io/projected/df758465-1b0a-4ecb-b70c-c256357e4ec0-kube-api-access-xq6kj\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.823074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-config\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.824143 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.824255 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-serving-cert\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.825062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-client\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.825285 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/daf6334f-8900-455a-95d7-8cc149fecc58-machine-approver-tls\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.825296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-trusted-ca\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.826298 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf6334f-8900-455a-95d7-8cc149fecc58-config\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.826362 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-service-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.826429 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f7190415-01a1-416c-b7a8-7945d9f20796-available-featuregates\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.826647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/13e04245-e686-4b2b-9b90-68659825e4ad-etcd-ca\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.827056 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.827315 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.827942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-serving-cert\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.828163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7190415-01a1-416c-b7a8-7945d9f20796-serving-cert\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.830010 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.830441 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.850431 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.870801 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.891296 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.915606 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.924838 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgjtz\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-kube-api-access-dgjtz\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.924904 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac8b263-5a8c-4621-a046-496c25f91286-serving-cert\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.924938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-proxy-tls\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.924967 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-images\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.924999 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925064 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925101 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e75817c-456c-4d5c-981d-c434adca9411-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925134 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925163 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838dd7d5-7935-4238-b000-4adb94d759c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925247 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2lmn\" (UniqueName: \"kubernetes.io/projected/704c868e-4e03-4945-9d88-382b59d8ffd1-kube-api-access-s2lmn\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925292 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925471 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-965z4\" (UniqueName: \"kubernetes.io/projected/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-kube-api-access-965z4\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925558 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq6kj\" (UniqueName: \"kubernetes.io/projected/df758465-1b0a-4ecb-b70c-c256357e4ec0-kube-api-access-xq6kj\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925617 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gndmp\" (UniqueName: \"kubernetes.io/projected/a398dbb8-67e9-4390-9a50-9c5dd2c62303-kube-api-access-gndmp\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925680 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrm78\" (UniqueName: \"kubernetes.io/projected/3b5aee17-20ec-4adb-ad67-a239878063a6-kube-api-access-nrm78\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925705 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-plugins-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e75817c-456c-4d5c-981d-c434adca9411-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925773 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj5sv\" (UniqueName: \"kubernetes.io/projected/3f3cf4c7-106d-4f21-9596-6bd2356af306-kube-api-access-gj5sv\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925809 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-config\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925835 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw8hl\" (UniqueName: \"kubernetes.io/projected/b8f35bb3-a509-49f1-a945-0a8170569c1d-kube-api-access-qw8hl\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925867 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8f35bb3-a509-49f1-a945-0a8170569c1d-proxy-tls\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925893 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925915 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/289092c0-4c05-4c6d-bb0f-d36cb003b23e-metrics-tls\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925962 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk5lh\" (UniqueName: \"kubernetes.io/projected/2cebb480-f94a-4fef-b128-281428436da2-kube-api-access-qk5lh\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.925980 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0061d337-ce1d-4925-b9d4-c6805aafe7ec-metrics-tls\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926015 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e98663-cc71-4af8-8a92-446851e11efe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926047 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngq6q\" (UniqueName: \"kubernetes.io/projected/0061d337-ce1d-4925-b9d4-c6805aafe7ec-kube-api-access-ngq6q\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926042 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-plugins-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926080 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926102 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9tvg\" (UniqueName: \"kubernetes.io/projected/bac8b263-5a8c-4621-a046-496c25f91286-kube-api-access-m9tvg\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838dd7d5-7935-4238-b000-4adb94d759c6-config\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-stats-auth\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926501 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsxrm\" (UniqueName: \"kubernetes.io/projected/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-kube-api-access-fsxrm\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926546 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-socket-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926584 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-metrics-certs\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926619 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-profile-collector-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926653 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx28x\" (UniqueName: \"kubernetes.io/projected/b04c7d45-8cef-4c51-8821-0eac0a6f6b07-kube-api-access-lx28x\") pod \"migrator-59844c95c7-j5tgf\" (UID: \"b04c7d45-8cef-4c51-8821-0eac0a6f6b07\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926670 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e75817c-456c-4d5c-981d-c434adca9411-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926697 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-service-ca-bundle\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926732 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926766 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5aee17-20ec-4adb-ad67-a239878063a6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.926912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838dd7d5-7935-4238-b000-4adb94d759c6-config\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-socket-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927048 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwfx9\" (UniqueName: \"kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927092 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzlxj\" (UniqueName: \"kubernetes.io/projected/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-kube-api-access-wzlxj\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927121 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a398dbb8-67e9-4390-9a50-9c5dd2c62303-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927149 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927173 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927220 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t945\" (UniqueName: \"kubernetes.io/projected/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-kube-api-access-2t945\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927283 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-default-certificate\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927395 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-mountpoint-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927452 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bv9\" (UniqueName: \"kubernetes.io/projected/60628569-ded6-49cb-8bb3-b814b5afc84a-kube-api-access-h2bv9\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927478 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927519 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-mountpoint-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927574 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kdwj\" (UniqueName: \"kubernetes.io/projected/d24f1f11-7f2c-4917-8648-cf1cc03d324b-kube-api-access-2kdwj\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927623 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-srv-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-csi-data-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/df758465-1b0a-4ecb-b70c-c256357e4ec0-tmpfs\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-csi-data-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.927983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrdhw\" (UniqueName: \"kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-srv-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928179 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/838dd7d5-7935-4238-b000-4adb94d759c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/df758465-1b0a-4ecb-b70c-c256357e4ec0-tmpfs\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928244 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928268 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db9df3df-527a-4573-b1ac-4bf2bd94444e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e98663-cc71-4af8-8a92-446851e11efe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grhwh\" (UniqueName: \"kubernetes.io/projected/67e98663-cc71-4af8-8a92-446851e11efe-kube-api-access-grhwh\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928458 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/289092c0-4c05-4c6d-bb0f-d36cb003b23e-trusted-ca\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cebb480-f94a-4fef-b128-281428436da2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928571 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkzrk\" (UniqueName: \"kubernetes.io/projected/db9df3df-527a-4573-b1ac-4bf2bd94444e-kube-api-access-rkzrk\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928620 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928658 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac8b263-5a8c-4621-a046-496c25f91286-config\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b5aee17-20ec-4adb-ad67-a239878063a6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928753 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-webhook-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928804 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-registration-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928849 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/60628569-ded6-49cb-8bb3-b814b5afc84a-registration-dir\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.928866 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e75817c-456c-4d5c-981d-c434adca9411-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.929332 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e75817c-456c-4d5c-981d-c434adca9411-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.931070 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.931178 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/838dd7d5-7935-4238-b000-4adb94d759c6-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.931773 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0061d337-ce1d-4925-b9d4-c6805aafe7ec-metrics-tls\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.932242 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/289092c0-4c05-4c6d-bb0f-d36cb003b23e-trusted-ca\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.932366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/289092c0-4c05-4c6d-bb0f-d36cb003b23e-metrics-tls\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.951038 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.969908 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 03 13:32:50 crc kubenswrapper[4959]: I1003 13:32:50.990768 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.005466 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.009833 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.020633 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b5aee17-20ec-4adb-ad67-a239878063a6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.034835 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.040165 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b5aee17-20ec-4adb-ad67-a239878063a6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.050089 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.070004 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.091173 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.109774 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.117725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e98663-cc71-4af8-8a92-446851e11efe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.130419 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.150041 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.162326 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e98663-cc71-4af8-8a92-446851e11efe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.170652 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.190296 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.211451 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.226254 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.230564 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.237505 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-config\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.269706 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.283933 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db9df3df-527a-4573-b1ac-4bf2bd94444e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.290100 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.310353 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.330222 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.349959 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.361564 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-default-certificate\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.370713 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.380471 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-stats-auth\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.389841 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.402268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-metrics-certs\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.410016 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.418957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-service-ca-bundle\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.430212 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.450918 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.457000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8f35bb3-a509-49f1-a945-0a8170569c1d-images\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.471486 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.479550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8f35bb3-a509-49f1-a945-0a8170569c1d-proxy-tls\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.491861 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.509812 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.522885 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cebb480-f94a-4fef-b128-281428436da2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.529609 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.539398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-proxy-tls\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.550478 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.570238 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.590128 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.599044 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.615664 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.621792 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.630009 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.650088 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.669765 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.690258 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.709180 4959 request.go:700] Waited for 1.013973617s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.711516 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.719493 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.720172 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-profile-collector-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.720247 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.729960 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.743131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3f3cf4c7-106d-4f21-9596-6bd2356af306-srv-cert\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.750914 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.770499 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.789982 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.801074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-srv-cert\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.810447 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.820840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a398dbb8-67e9-4390-9a50-9c5dd2c62303-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.830706 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.839083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.842126 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df758465-1b0a-4ecb-b70c-c256357e4ec0-webhook-cert\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.851080 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.859363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bac8b263-5a8c-4621-a046-496c25f91286-serving-cert\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.870255 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.879736 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bac8b263-5a8c-4621-a046-496c25f91286-config\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.890413 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.910477 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.925784 4959 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.925827 4959 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.925869 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle podName:d24f1f11-7f2c-4917-8648-cf1cc03d324b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.42584799 +0000 UTC m=+141.629191407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle") pod "service-ca-9c57cc56f-l2prt" (UID: "d24f1f11-7f2c-4917-8648-cf1cc03d324b") : failed to sync configmap cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.925918 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert podName:0d0f85d6-3dec-4c08-8087-4abab2a3ab53 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.425893562 +0000 UTC m=+141.629236979 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert") pod "ingress-canary-452zl" (UID: "0d0f85d6-3dec-4c08-8087-4abab2a3ab53") : failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.927037 4959 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.927383 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token podName:704c868e-4e03-4945-9d88-382b59d8ffd1 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.427281398 +0000 UTC m=+141.630624845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token") pod "machine-config-server-bg4hd" (UID: "704c868e-4e03-4945-9d88-382b59d8ffd1") : failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.927428 4959 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.927522 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key podName:d24f1f11-7f2c-4917-8648-cf1cc03d324b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.427505375 +0000 UTC m=+141.630848832 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key") pod "service-ca-9c57cc56f-l2prt" (UID: "d24f1f11-7f2c-4917-8648-cf1cc03d324b") : failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.927990 4959 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.929186 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume podName:c6c97e09-88e1-4f2b-a476-19796c3d83f2 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.428039699 +0000 UTC m=+141.631383196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume") pod "collect-profiles-29324970-2gdjd" (UID: "c6c97e09-88e1-4f2b-a476-19796c3d83f2") : failed to sync configmap cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.929300 4959 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: E1003 13:32:51.929404 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs podName:704c868e-4e03-4945-9d88-382b59d8ffd1 nodeName:}" failed. No retries permitted until 2025-10-03 13:32:52.429383954 +0000 UTC m=+141.632727441 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs") pod "machine-config-server-bg4hd" (UID: "704c868e-4e03-4945-9d88-382b59d8ffd1") : failed to sync secret cache: timed out waiting for the condition Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.929994 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.949835 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.970028 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 03 13:32:51 crc kubenswrapper[4959]: I1003 13:32:51.990862 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.010500 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.030501 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.075046 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggp4d\" (UniqueName: \"kubernetes.io/projected/a7ebf605-04c0-41ab-bf73-2c0ae57be8a6-kube-api-access-ggp4d\") pod \"openshift-apiserver-operator-796bbdcf4f-d8q9d\" (UID: \"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.087377 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4fg\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-kube-api-access-7w4fg\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.109153 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df5c87b-62df-4f92-a7e0-3f06aa2aad08-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rmq68\" (UID: \"2df5c87b-62df-4f92-a7e0-3f06aa2aad08\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.110114 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.130949 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.150531 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.170160 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.190491 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.223396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c24kn\" (UniqueName: \"kubernetes.io/projected/677b5491-6c28-4625-854b-881e25662a57-kube-api-access-c24kn\") pod \"apiserver-76f77b778f-h7s26\" (UID: \"677b5491-6c28-4625-854b-881e25662a57\") " pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.243949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6k7d\" (UniqueName: \"kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d\") pod \"route-controller-manager-6576b87f9c-9zvr5\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.263237 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftqb4\" (UniqueName: \"kubernetes.io/projected/5cdb1c45-cb05-4d7d-9d75-693e90227102-kube-api-access-ftqb4\") pod \"cluster-samples-operator-665b6dd947-m5j8l\" (UID: \"5cdb1c45-cb05-4d7d-9d75-693e90227102\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.271382 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.285425 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.286559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6258\" (UniqueName: \"kubernetes.io/projected/5c097a19-754b-465c-8c78-fc8b9ea45526-kube-api-access-r6258\") pod \"apiserver-7bbb656c7d-5l9tb\" (UID: \"5c097a19-754b-465c-8c78-fc8b9ea45526\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.290570 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.301082 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.310386 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.322452 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.330556 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.336022 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.348731 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.373626 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w7pt\" (UniqueName: \"kubernetes.io/projected/fcf06d86-50e8-4978-81ad-d9a2351afc39-kube-api-access-5w7pt\") pod \"authentication-operator-69f744f599-s7kpd\" (UID: \"fcf06d86-50e8-4978-81ad-d9a2351afc39\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.398026 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pwqb\" (UniqueName: \"kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb\") pod \"oauth-openshift-558db77b4-jqpvt\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.407277 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpxrr\" (UniqueName: \"kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr\") pod \"controller-manager-879f6c89f-b5jrs\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.432937 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l67h9\" (UniqueName: \"kubernetes.io/projected/02ca8bcb-1ea4-4281-a85d-e74eb179aa39-kube-api-access-l67h9\") pod \"machine-api-operator-5694c8668f-n9cfz\" (UID: \"02ca8bcb-1ea4-4281-a85d-e74eb179aa39\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.438602 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.451568 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460693 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460755 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460802 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.460874 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.461860 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-cabundle\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.461864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.465899 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-certs\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.466181 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/d24f1f11-7f2c-4917-8648-cf1cc03d324b-signing-key\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.470409 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.473765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/704c868e-4e03-4945-9d88-382b59d8ffd1-node-bootstrap-token\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.490819 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.509909 4959 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.530966 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.538340 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.549785 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.550160 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.570077 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.574248 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-cert\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.590225 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.610772 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.610907 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.629918 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.653381 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.670099 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.688485 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.705064 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.709986 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-n9cfz"] Oct 03 13:32:52 crc kubenswrapper[4959]: W1003 13:32:52.720288 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02ca8bcb_1ea4_4281_a85d_e74eb179aa39.slice/crio-5737177778f47571b83a3aef02e2f29e874a558741f2b75a644e12b3ab06b49a WatchSource:0}: Error finding container 5737177778f47571b83a3aef02e2f29e874a558741f2b75a644e12b3ab06b49a: Status 404 returned error can't find the container with id 5737177778f47571b83a3aef02e2f29e874a558741f2b75a644e12b3ab06b49a Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.727944 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h7s26"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.728388 4959 request.go:700] Waited for 1.9060467s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.736775 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xv6l\" (UniqueName: \"kubernetes.io/projected/13e04245-e686-4b2b-9b90-68659825e4ad-kube-api-access-9xv6l\") pod \"etcd-operator-b45778765-z6rtc\" (UID: \"13e04245-e686-4b2b-9b90-68659825e4ad\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.746949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.753887 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6q67\" (UniqueName: \"kubernetes.io/projected/daf6334f-8900-455a-95d7-8cc149fecc58-kube-api-access-w6q67\") pod \"machine-approver-56656f9798-xn7cr\" (UID: \"daf6334f-8900-455a-95d7-8cc149fecc58\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.764577 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.772363 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.782578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6mc\" (UniqueName: \"kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc\") pod \"console-f9d7485db-l2fnx\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.797925 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl2p6\" (UniqueName: \"kubernetes.io/projected/7d1875ce-7ea7-47ee-a495-d1d51c0b56e4-kube-api-access-nl2p6\") pod \"console-operator-58897d9998-fzwkw\" (UID: \"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4\") " pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.803541 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.820829 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96gx2\" (UniqueName: \"kubernetes.io/projected/f7190415-01a1-416c-b7a8-7945d9f20796-kube-api-access-96gx2\") pod \"openshift-config-operator-7777fb866f-nqgtl\" (UID: \"f7190415-01a1-416c-b7a8-7945d9f20796\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.826427 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.827012 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zcp5\" (UniqueName: \"kubernetes.io/projected/847fd56c-ec05-4627-bd43-49add0bdc748-kube-api-access-6zcp5\") pod \"downloads-7954f5f757-ttm4r\" (UID: \"847fd56c-ec05-4627-bd43-49add0bdc748\") " pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.829996 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.833007 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s7kpd"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.846930 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgjtz\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-kube-api-access-dgjtz\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:52 crc kubenswrapper[4959]: W1003 13:32:52.851206 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7ebf605_04c0_41ab_bf73_2c0ae57be8a6.slice/crio-69ee3e40d212a50e9f8639af6191e8f7db629269a36aa76e195562690771ce2e WatchSource:0}: Error finding container 69ee3e40d212a50e9f8639af6191e8f7db629269a36aa76e195562690771ce2e: Status 404 returned error can't find the container with id 69ee3e40d212a50e9f8639af6191e8f7db629269a36aa76e195562690771ce2e Oct 03 13:32:52 crc kubenswrapper[4959]: W1003 13:32:52.863726 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c097a19_754b_465c_8c78_fc8b9ea45526.slice/crio-ea323ae18360d14ad40b81da7993b96f5614283372af8246ee5adcf41d11b5e0 WatchSource:0}: Error finding container ea323ae18360d14ad40b81da7993b96f5614283372af8246ee5adcf41d11b5e0: Status 404 returned error can't find the container with id ea323ae18360d14ad40b81da7993b96f5614283372af8246ee5adcf41d11b5e0 Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.878965 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2lmn\" (UniqueName: \"kubernetes.io/projected/704c868e-4e03-4945-9d88-382b59d8ffd1-kube-api-access-s2lmn\") pod \"machine-config-server-bg4hd\" (UID: \"704c868e-4e03-4945-9d88-382b59d8ffd1\") " pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.886379 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-965z4\" (UniqueName: \"kubernetes.io/projected/3393e55a-a44e-42d8-a1ad-7c62ead0edc2-kube-api-access-965z4\") pod \"router-default-5444994796-k89zn\" (UID: \"3393e55a-a44e-42d8-a1ad-7c62ead0edc2\") " pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.891702 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.914677 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.917439 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq6kj\" (UniqueName: \"kubernetes.io/projected/df758465-1b0a-4ecb-b70c-c256357e4ec0-kube-api-access-xq6kj\") pod \"packageserver-d55dfcdfc-fpgj7\" (UID: \"df758465-1b0a-4ecb-b70c-c256357e4ec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.938042 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrm78\" (UniqueName: \"kubernetes.io/projected/3b5aee17-20ec-4adb-ad67-a239878063a6-kube-api-access-nrm78\") pod \"openshift-controller-manager-operator-756b6f6bc6-ff9j6\" (UID: \"3b5aee17-20ec-4adb-ad67-a239878063a6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.946913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj5sv\" (UniqueName: \"kubernetes.io/projected/3f3cf4c7-106d-4f21-9596-6bd2356af306-kube-api-access-gj5sv\") pod \"catalog-operator-68c6474976-s9jrh\" (UID: \"3f3cf4c7-106d-4f21-9596-6bd2356af306\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.951110 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:52 crc kubenswrapper[4959]: W1003 13:32:52.951955 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa273389_8a40_4e29_b0cd_1be72ac79cb6.slice/crio-3150e663e7375ed21a6be27f04aa1a4b050b80efcb193a009f90ed3dd0690acb WatchSource:0}: Error finding container 3150e663e7375ed21a6be27f04aa1a4b050b80efcb193a009f90ed3dd0690acb: Status 404 returned error can't find the container with id 3150e663e7375ed21a6be27f04aa1a4b050b80efcb193a009f90ed3dd0690acb Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.972510 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.973285 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw8hl\" (UniqueName: \"kubernetes.io/projected/b8f35bb3-a509-49f1-a945-0a8170569c1d-kube-api-access-qw8hl\") pod \"machine-config-operator-74547568cd-cg2fr\" (UID: \"b8f35bb3-a509-49f1-a945-0a8170569c1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:52 crc kubenswrapper[4959]: I1003 13:32:52.983872 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gndmp\" (UniqueName: \"kubernetes.io/projected/a398dbb8-67e9-4390-9a50-9c5dd2c62303-kube-api-access-gndmp\") pod \"package-server-manager-789f6589d5-9h4nn\" (UID: \"a398dbb8-67e9-4390-9a50-9c5dd2c62303\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:52 crc kubenswrapper[4959]: W1003 13:32:52.987758 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3393e55a_a44e_42d8_a1ad_7c62ead0edc2.slice/crio-97f38b9016a6565d9ff412761e0e4c93573de5296ec672992293a7ad4d3582cc WatchSource:0}: Error finding container 97f38b9016a6565d9ff412761e0e4c93573de5296ec672992293a7ad4d3582cc: Status 404 returned error can't find the container with id 97f38b9016a6565d9ff412761e0e4c93573de5296ec672992293a7ad4d3582cc Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.009321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9tvg\" (UniqueName: \"kubernetes.io/projected/bac8b263-5a8c-4621-a046-496c25f91286-kube-api-access-m9tvg\") pod \"service-ca-operator-777779d784-h4pbx\" (UID: \"bac8b263-5a8c-4621-a046-496c25f91286\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.017914 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.029095 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk5lh\" (UniqueName: \"kubernetes.io/projected/2cebb480-f94a-4fef-b128-281428436da2-kube-api-access-qk5lh\") pod \"multus-admission-controller-857f4d67dd-v6d5j\" (UID: \"2cebb480-f94a-4fef-b128-281428436da2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.037456 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bg4hd" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.058101 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.059204 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsxrm\" (UniqueName: \"kubernetes.io/projected/50e95aa6-57b0-4b4d-9d6c-c48c129c1052-kube-api-access-fsxrm\") pod \"olm-operator-6b444d44fb-bcgsg\" (UID: \"50e95aa6-57b0-4b4d-9d6c-c48c129c1052\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.062000 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z6rtc"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.069565 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.077228 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.083501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx28x\" (UniqueName: \"kubernetes.io/projected/b04c7d45-8cef-4c51-8821-0eac0a6f6b07-kube-api-access-lx28x\") pod \"migrator-59844c95c7-j5tgf\" (UID: \"b04c7d45-8cef-4c51-8821-0eac0a6f6b07\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.086060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.091599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzlxj\" (UniqueName: \"kubernetes.io/projected/d3845ba1-19b1-4a7b-996c-ca26ffd8cda0-kube-api-access-wzlxj\") pod \"machine-config-controller-84d6567774-z2kjq\" (UID: \"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.106366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwfx9\" (UniqueName: \"kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9\") pod \"collect-profiles-29324970-2gdjd\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.133574 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngq6q\" (UniqueName: \"kubernetes.io/projected/0061d337-ce1d-4925-b9d4-c6805aafe7ec-kube-api-access-ngq6q\") pod \"dns-operator-744455d44c-zsxvg\" (UID: \"0061d337-ce1d-4925-b9d4-c6805aafe7ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:53 crc kubenswrapper[4959]: W1003 13:32:53.140293 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaf6334f_8900_455a_95d7_8cc149fecc58.slice/crio-ee015279afb8dca46ccfdc1fa86f116e259b494e2ad595da64296a52f8a6631e WatchSource:0}: Error finding container ee015279afb8dca46ccfdc1fa86f116e259b494e2ad595da64296a52f8a6631e: Status 404 returned error can't find the container with id ee015279afb8dca46ccfdc1fa86f116e259b494e2ad595da64296a52f8a6631e Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.154599 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.159146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t945\" (UniqueName: \"kubernetes.io/projected/0d0f85d6-3dec-4c08-8087-4abab2a3ab53-kube-api-access-2t945\") pod \"ingress-canary-452zl\" (UID: \"0d0f85d6-3dec-4c08-8087-4abab2a3ab53\") " pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.167087 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.171358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bv9\" (UniqueName: \"kubernetes.io/projected/60628569-ded6-49cb-8bb3-b814b5afc84a-kube-api-access-h2bv9\") pod \"csi-hostpathplugin-v8x2x\" (UID: \"60628569-ded6-49cb-8bb3-b814b5afc84a\") " pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.194409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/289092c0-4c05-4c6d-bb0f-d36cb003b23e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vms5l\" (UID: \"289092c0-4c05-4c6d-bb0f-d36cb003b23e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:53 crc kubenswrapper[4959]: W1003 13:32:53.195131 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod704c868e_4e03_4945_9d88_382b59d8ffd1.slice/crio-b6295c971055702f17192e8371e4c76a880c638b3965316beccdf2691c69afa1 WatchSource:0}: Error finding container b6295c971055702f17192e8371e4c76a880c638b3965316beccdf2691c69afa1: Status 404 returned error can't find the container with id b6295c971055702f17192e8371e4c76a880c638b3965316beccdf2691c69afa1 Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.199085 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.206553 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.210752 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kdwj\" (UniqueName: \"kubernetes.io/projected/d24f1f11-7f2c-4917-8648-cf1cc03d324b-kube-api-access-2kdwj\") pod \"service-ca-9c57cc56f-l2prt\" (UID: \"d24f1f11-7f2c-4917-8648-cf1cc03d324b\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.214402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.227016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrdhw\" (UniqueName: \"kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw\") pod \"marketplace-operator-79b997595-gspcs\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.233729 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.248299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/838dd7d5-7935-4238-b000-4adb94d759c6-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wlc98\" (UID: \"838dd7d5-7935-4238-b000-4adb94d759c6\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.257507 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.264493 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.266616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grhwh\" (UniqueName: \"kubernetes.io/projected/67e98663-cc71-4af8-8a92-446851e11efe-kube-api-access-grhwh\") pod \"kube-storage-version-migrator-operator-b67b599dd-thbnm\" (UID: \"67e98663-cc71-4af8-8a92-446851e11efe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.271159 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.281025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.291703 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.305367 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.307460 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/275d0dcc-f61b-4ffe-a19e-bc3a2e996861-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f74nt\" (UID: \"275d0dcc-f61b-4ffe-a19e-bc3a2e996861\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.328502 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.333544 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkzrk\" (UniqueName: \"kubernetes.io/projected/db9df3df-527a-4573-b1ac-4bf2bd94444e-kube-api-access-rkzrk\") pod \"control-plane-machine-set-operator-78cbb6b69f-9l2d2\" (UID: \"db9df3df-527a-4573-b1ac-4bf2bd94444e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.352767 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e75817c-456c-4d5c-981d-c434adca9411-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-47mhg\" (UID: \"2e75817c-456c-4d5c-981d-c434adca9411\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.367279 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.368696 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" event={"ID":"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b","Type":"ContainerStarted","Data":"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.368758 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" event={"ID":"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b","Type":"ContainerStarted","Data":"720ed1c7f3f0e1d8c22f04da443461f7bd89dc88f4a853c876e2b944ff12505e"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.369880 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.373963 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-452zl" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.380168 4959 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-b5jrs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.380242 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392222 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4qfh\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392278 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/132cc225-bf51-49a1-8a9d-0058984db61e-config-volume\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392321 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392455 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/132cc225-bf51-49a1-8a9d-0058984db61e-metrics-tls\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392492 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392540 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.392654 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzfl9\" (UniqueName: \"kubernetes.io/projected/132cc225-bf51-49a1-8a9d-0058984db61e-kube-api-access-gzfl9\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.393062 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:53.893046752 +0000 UTC m=+143.096390169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.398509 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.403871 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.412170 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.414230 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" event={"ID":"3f3cf4c7-106d-4f21-9596-6bd2356af306","Type":"ContainerStarted","Data":"9690acc1cd662168baae4ca7bc69d477d848c42a452fcc6bb2b42f2294580b04"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.445133 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" event={"ID":"5cdb1c45-cb05-4d7d-9d75-693e90227102","Type":"ContainerStarted","Data":"50702ffa1572847c3f6faa2dea0e6255ed4faec72c6fdb75f8529bdca2dceb7c"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.450408 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" event={"ID":"2df5c87b-62df-4f92-a7e0-3f06aa2aad08","Type":"ContainerStarted","Data":"53fd1e986bb1db28dff9f17cd7154ac6ae892b340e677e838b7a8a1cad13e7fd"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.450452 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" event={"ID":"2df5c87b-62df-4f92-a7e0-3f06aa2aad08","Type":"ContainerStarted","Data":"eb11b2f6317b9538e3c97bddef367bc287bbad4d06f9eb08d30c0a72901bca65"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.458915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" event={"ID":"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6","Type":"ContainerStarted","Data":"9a80cc40e0d9b5f0741e1f8d46b0fd650f549d959f534d0df7085887cb170d05"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.459222 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" event={"ID":"a7ebf605-04c0-41ab-bf73-2c0ae57be8a6","Type":"ContainerStarted","Data":"69ee3e40d212a50e9f8639af6191e8f7db629269a36aa76e195562690771ce2e"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.460404 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.468491 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.469960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" event={"ID":"aa273389-8a40-4e29-b0cd-1be72ac79cb6","Type":"ContainerStarted","Data":"3150e663e7375ed21a6be27f04aa1a4b050b80efcb193a009f90ed3dd0690acb"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.471120 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bg4hd" event={"ID":"704c868e-4e03-4945-9d88-382b59d8ffd1","Type":"ContainerStarted","Data":"b6295c971055702f17192e8371e4c76a880c638b3965316beccdf2691c69afa1"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.475310 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.478982 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k89zn" event={"ID":"3393e55a-a44e-42d8-a1ad-7c62ead0edc2","Type":"ContainerStarted","Data":"97f38b9016a6565d9ff412761e0e4c93573de5296ec672992293a7ad4d3582cc"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.481834 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" event={"ID":"5c097a19-754b-465c-8c78-fc8b9ea45526","Type":"ContainerStarted","Data":"ea323ae18360d14ad40b81da7993b96f5614283372af8246ee5adcf41d11b5e0"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.482027 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.490689 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" event={"ID":"02ca8bcb-1ea4-4281-a85d-e74eb179aa39","Type":"ContainerStarted","Data":"7feb991265eae1b6c1eb6178267006d263788c918823f858cc2896f9efc023b3"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.490763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" event={"ID":"02ca8bcb-1ea4-4281-a85d-e74eb179aa39","Type":"ContainerStarted","Data":"5737177778f47571b83a3aef02e2f29e874a558741f2b75a644e12b3ab06b49a"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.492402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" event={"ID":"fcf06d86-50e8-4978-81ad-d9a2351afc39","Type":"ContainerStarted","Data":"9cbd9ab8200c2dc5535efc77c19c20ff953d5a55ab8a37208bb81a58f3b542e6"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.492442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" event={"ID":"fcf06d86-50e8-4978-81ad-d9a2351afc39","Type":"ContainerStarted","Data":"e4677a11e36d8d7d828cee0a3ec9c9f4a3e281c92357405a253157ee6f635fea"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.493104 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.493360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.493457 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.493538 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzfl9\" (UniqueName: \"kubernetes.io/projected/132cc225-bf51-49a1-8a9d-0058984db61e-kube-api-access-gzfl9\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.494459 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.494543 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:53.99452645 +0000 UTC m=+143.197869867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4qfh\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495465 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/132cc225-bf51-49a1-8a9d-0058984db61e-config-volume\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495538 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495657 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495709 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495764 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/132cc225-bf51-49a1-8a9d-0058984db61e-metrics-tls\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.495816 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.502184 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/132cc225-bf51-49a1-8a9d-0058984db61e-config-volume\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.502839 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" event={"ID":"daf6334f-8900-455a-95d7-8cc149fecc58","Type":"ContainerStarted","Data":"ee015279afb8dca46ccfdc1fa86f116e259b494e2ad595da64296a52f8a6631e"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.503134 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.504223 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.504509 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.004478092 +0000 UTC m=+143.207821589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.504533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.504757 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.510671 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/132cc225-bf51-49a1-8a9d-0058984db61e-metrics-tls\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.512548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" event={"ID":"df758465-1b0a-4ecb-b70c-c256357e4ec0","Type":"ContainerStarted","Data":"7c09611accb5709ef986e70b0c3a6d3de939f3c4749314264150e603e655ced0"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.520549 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" event={"ID":"13e04245-e686-4b2b-9b90-68659825e4ad","Type":"ContainerStarted","Data":"61cfd641892bb7a4987f67deea612697760a5d1eb04155f44221a6c36bf627b7"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.521926 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.524111 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.524517 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" event={"ID":"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac","Type":"ContainerStarted","Data":"957e53ca5689590cbe482d040611f2f1326aa04d89d60ac91af36ada86e6a393"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.524561 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" event={"ID":"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac","Type":"ContainerStarted","Data":"aad9367e69cb42da54e93991aa974bc23fc61905c47fb3e30b7f9e298561dcae"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.525289 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.532732 4959 generic.go:334] "Generic (PLEG): container finished" podID="677b5491-6c28-4625-854b-881e25662a57" containerID="4105fb5b819088793b3e21f25354f67340b3af513d4d7bfe5370d57ea1beec80" exitCode=0 Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.532794 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" event={"ID":"677b5491-6c28-4625-854b-881e25662a57","Type":"ContainerDied","Data":"4105fb5b819088793b3e21f25354f67340b3af513d4d7bfe5370d57ea1beec80"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.532829 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" event={"ID":"677b5491-6c28-4625-854b-881e25662a57","Type":"ContainerStarted","Data":"d9ce9984f79652525ca987b483e446e233cec218d151894983d01544722b4822"} Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.534967 4959 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-9zvr5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.535014 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.539240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzfl9\" (UniqueName: \"kubernetes.io/projected/132cc225-bf51-49a1-8a9d-0058984db61e-kube-api-access-gzfl9\") pod \"dns-default-mw5qw\" (UID: \"132cc225-bf51-49a1-8a9d-0058984db61e\") " pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.549210 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4qfh\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.549386 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ttm4r"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.587577 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-fzwkw"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.593699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.596461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.598998 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.098964505 +0000 UTC m=+143.302307952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.679577 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.698380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.698662 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.198649497 +0000 UTC m=+143.401992914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.799663 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.800236 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.300211486 +0000 UTC m=+143.503554903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.887874 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6"] Oct 03 13:32:53 crc kubenswrapper[4959]: I1003 13:32:53.902005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:53 crc kubenswrapper[4959]: E1003 13:32:53.902334 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.402321961 +0000 UTC m=+143.605665378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.006019 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.006688 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.506666474 +0000 UTC m=+143.710009891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.044142 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq"] Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.091543 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zsxvg"] Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.108311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.108727 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.608704906 +0000 UTC m=+143.812048323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.208909 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.209245 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.709225979 +0000 UTC m=+143.912569396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: W1003 13:32:54.242332 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3845ba1_19b1_4a7b_996c_ca26ffd8cda0.slice/crio-543c37b4152f64a3a0ea5e552dbc0385fe20d280f7e3c8e748014f3798860197 WatchSource:0}: Error finding container 543c37b4152f64a3a0ea5e552dbc0385fe20d280f7e3c8e748014f3798860197: Status 404 returned error can't find the container with id 543c37b4152f64a3a0ea5e552dbc0385fe20d280f7e3c8e748014f3798860197 Oct 03 13:32:54 crc kubenswrapper[4959]: W1003 13:32:54.254351 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0061d337_ce1d_4925_b9d4_c6805aafe7ec.slice/crio-91f4bf4b18b65baaa7f42450841605146cecad80f28d0fd0afa2fc2f3d12c064 WatchSource:0}: Error finding container 91f4bf4b18b65baaa7f42450841605146cecad80f28d0fd0afa2fc2f3d12c064: Status 404 returned error can't find the container with id 91f4bf4b18b65baaa7f42450841605146cecad80f28d0fd0afa2fc2f3d12c064 Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.311091 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.311785 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.811770195 +0000 UTC m=+144.015113612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.387997 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg"] Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.412355 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.412820 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:54.912795141 +0000 UTC m=+144.116138558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.515049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.515469 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.015451399 +0000 UTC m=+144.218794826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.562972 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr"] Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.606141 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" event={"ID":"f7190415-01a1-416c-b7a8-7945d9f20796","Type":"ContainerStarted","Data":"f154905728a3857c14d2271030f5cf642ee2247f0bb4cbefe0cfcefbdff778ea"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.624422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.624840 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.124788464 +0000 UTC m=+144.328131881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.625123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.638819 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.138793832 +0000 UTC m=+144.342137249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.651111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" event={"ID":"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0","Type":"ContainerStarted","Data":"543c37b4152f64a3a0ea5e552dbc0385fe20d280f7e3c8e748014f3798860197"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.653690 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98"] Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.708620 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" event={"ID":"13e04245-e686-4b2b-9b90-68659825e4ad","Type":"ContainerStarted","Data":"956ca9c94e92faec1702b4aee5859b241ab14143461fd83bb2b9a68d1d2aa907"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.728594 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.729219 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.229180009 +0000 UTC m=+144.432523426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.732255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ttm4r" event={"ID":"847fd56c-ec05-4627-bd43-49add0bdc748","Type":"ContainerStarted","Data":"294f2a73e1b7e7c83049b934b8e1856915d87d72d25c28979f400e4d74d07c24"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.814905 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" podStartSLOduration=123.814879381 podStartE2EDuration="2m3.814879381s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:54.759905036 +0000 UTC m=+143.963248453" watchObservedRunningTime="2025-10-03 13:32:54.814879381 +0000 UTC m=+144.018222798" Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.831059 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.832268 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.332251198 +0000 UTC m=+144.535594615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.901830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" event={"ID":"3b5aee17-20ec-4adb-ad67-a239878063a6","Type":"ContainerStarted","Data":"340d0c3f8aea79b53ae2ac697c00bd62214f9516f15bd093321d72a43dcffb6b"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.933500 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:54 crc kubenswrapper[4959]: E1003 13:32:54.933922 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.433895 +0000 UTC m=+144.637238427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.937276 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" event={"ID":"02ca8bcb-1ea4-4281-a85d-e74eb179aa39","Type":"ContainerStarted","Data":"459d07ceec715b09c711db662267aaf0b5ebca311187e44a295d19d78f08acaa"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.958047 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" event={"ID":"daf6334f-8900-455a-95d7-8cc149fecc58","Type":"ContainerStarted","Data":"c5cb71aa6e0e39b76f132503dc2a17ef10057b70e3485bb96b8029c37adbf50a"} Oct 03 13:32:54 crc kubenswrapper[4959]: I1003 13:32:54.972808 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-n9cfz" podStartSLOduration=123.972774692 podStartE2EDuration="2m3.972774692s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:54.932851662 +0000 UTC m=+144.136195089" watchObservedRunningTime="2025-10-03 13:32:54.972774692 +0000 UTC m=+144.176118109" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.019118 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l2fnx" event={"ID":"819c5b6d-500f-4954-8225-a229c415bca9","Type":"ContainerStarted","Data":"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.019171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l2fnx" event={"ID":"819c5b6d-500f-4954-8225-a229c415bca9","Type":"ContainerStarted","Data":"2d694048f15b02f41a4e83e23e00c360fb33a6e412dcfe58259275a95189a09e"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.034271 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" event={"ID":"aa273389-8a40-4e29-b0cd-1be72ac79cb6","Type":"ContainerStarted","Data":"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.035485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.039623 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.539582158 +0000 UTC m=+144.742925575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.048094 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rmq68" podStartSLOduration=124.048068501 podStartE2EDuration="2m4.048068501s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.044116798 +0000 UTC m=+144.247460215" watchObservedRunningTime="2025-10-03 13:32:55.048068501 +0000 UTC m=+144.251411928" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.048607 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.069255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" event={"ID":"0061d337-ce1d-4925-b9d4-c6805aafe7ec","Type":"ContainerStarted","Data":"91f4bf4b18b65baaa7f42450841605146cecad80f28d0fd0afa2fc2f3d12c064"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.105695 4959 generic.go:334] "Generic (PLEG): container finished" podID="5c097a19-754b-465c-8c78-fc8b9ea45526" containerID="c7f6b7855ca974e340bc158751d343550014a209879eb43aa16bd6e997b1a9a1" exitCode=0 Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.105808 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" event={"ID":"5c097a19-754b-465c-8c78-fc8b9ea45526","Type":"ContainerDied","Data":"c7f6b7855ca974e340bc158751d343550014a209879eb43aa16bd6e997b1a9a1"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.126600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" event={"ID":"5cdb1c45-cb05-4d7d-9d75-693e90227102","Type":"ContainerStarted","Data":"8603c3f149cff7c564b03e3b377f5a92e3791c839962855b545b0d560f562af4"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.129173 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d8q9d" podStartSLOduration=125.129149923 podStartE2EDuration="2m5.129149923s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.126219346 +0000 UTC m=+144.329562763" watchObservedRunningTime="2025-10-03 13:32:55.129149923 +0000 UTC m=+144.332493340" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.136599 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.137963 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.637944645 +0000 UTC m=+144.841288052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.141149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-k89zn" event={"ID":"3393e55a-a44e-42d8-a1ad-7c62ead0edc2","Type":"ContainerStarted","Data":"804b45d37361e1c834cd7a4634888d946d5cf04f6792dbb3b956d0e7742bb7e9"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.149540 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" event={"ID":"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4","Type":"ContainerStarted","Data":"8953b65677dd75daca63dbc9211d46cb5bab0961922f50f56048052ebaea97d1"} Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.150666 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.160524 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.161455 4959 patch_prober.go:28] interesting pod/console-operator-58897d9998-fzwkw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.161511 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" podUID="7d1875ce-7ea7-47ee-a495-d1d51c0b56e4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.163031 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.235078 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s7kpd" podStartSLOduration=125.235061217 podStartE2EDuration="2m5.235061217s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.183626185 +0000 UTC m=+144.386969602" watchObservedRunningTime="2025-10-03 13:32:55.235061217 +0000 UTC m=+144.438404644" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.238748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.242000 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.741983909 +0000 UTC m=+144.945327326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.273390 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" podStartSLOduration=124.273371874 podStartE2EDuration="2m4.273371874s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.268292771 +0000 UTC m=+144.471636198" watchObservedRunningTime="2025-10-03 13:32:55.273371874 +0000 UTC m=+144.476715291" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.340080 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.340310 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.840281133 +0000 UTC m=+145.043624550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.340916 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.349059 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.849039183 +0000 UTC m=+145.052382600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.404526 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" podStartSLOduration=124.404507462 podStartE2EDuration="2m4.404507462s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.403159166 +0000 UTC m=+144.606502593" watchObservedRunningTime="2025-10-03 13:32:55.404507462 +0000 UTC m=+144.607850879" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.462801 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.464652 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" podStartSLOduration=124.464623372 podStartE2EDuration="2m4.464623372s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.458246344 +0000 UTC m=+144.661589791" watchObservedRunningTime="2025-10-03 13:32:55.464623372 +0000 UTC m=+144.667966789" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.465179 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.964961061 +0000 UTC m=+145.168304478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.472287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.472835 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:55.972819408 +0000 UTC m=+145.176162835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.554614 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-k89zn" podStartSLOduration=124.554576077 podStartE2EDuration="2m4.554576077s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.546092464 +0000 UTC m=+144.749435891" watchObservedRunningTime="2025-10-03 13:32:55.554576077 +0000 UTC m=+144.757919494" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.574264 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.574604 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.074575182 +0000 UTC m=+145.277918599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.576503 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" podStartSLOduration=125.576490283 podStartE2EDuration="2m5.576490283s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.574318216 +0000 UTC m=+144.777661633" watchObservedRunningTime="2025-10-03 13:32:55.576490283 +0000 UTC m=+144.779833700" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.617248 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-v6d5j"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.618590 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" podStartSLOduration=125.61856943 podStartE2EDuration="2m5.61856943s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.606285076 +0000 UTC m=+144.809628513" watchObservedRunningTime="2025-10-03 13:32:55.61856943 +0000 UTC m=+144.821912847" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.656821 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.657959 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-l2fnx" podStartSLOduration=124.657931764 podStartE2EDuration="2m4.657931764s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.647355996 +0000 UTC m=+144.850699423" watchObservedRunningTime="2025-10-03 13:32:55.657931764 +0000 UTC m=+144.861275181" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.676635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.677618 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.17757153 +0000 UTC m=+145.380914947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.746133 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-z6rtc" podStartSLOduration=124.746095452 podStartE2EDuration="2m4.746095452s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:55.743810011 +0000 UTC m=+144.947153428" watchObservedRunningTime="2025-10-03 13:32:55.746095452 +0000 UTC m=+144.949438869" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.747385 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.747510 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2prt"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.779112 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.779735 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.279714776 +0000 UTC m=+145.483058193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.880913 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.881284 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.381272965 +0000 UTC m=+145.584616382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.884827 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.894755 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.926771 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:32:55 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:32:55 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:32:55 crc kubenswrapper[4959]: healthz check failed Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.926819 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.945460 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l"] Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.959149 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:32:55 crc kubenswrapper[4959]: I1003 13:32:55.986653 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:55 crc kubenswrapper[4959]: E1003 13:32:55.987298 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.487262642 +0000 UTC m=+145.690606049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.005783 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.073312 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.095054 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.101730 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.601711631 +0000 UTC m=+145.805055048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.172380 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-452zl"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.197976 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.221388 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.222180 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.722159097 +0000 UTC m=+145.925502514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.239745 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" event={"ID":"838dd7d5-7935-4238-b000-4adb94d759c6","Type":"ContainerStarted","Data":"0ff1841dbd4b073cea82e006f3515978bce6961f396eac6269ce6a95ff5cec26"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.239791 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" event={"ID":"838dd7d5-7935-4238-b000-4adb94d759c6","Type":"ContainerStarted","Data":"4e4a12d4cc55c3519f38243ad0de9f648a7356a160d10cceac304fed84f94d19"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.253269 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.253804 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.265395 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-v8x2x"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.283098 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-m5j8l" event={"ID":"5cdb1c45-cb05-4d7d-9d75-693e90227102","Type":"ContainerStarted","Data":"fc79fbb9b29f82e0e05b2cc7cea0a529825548644bf7a57d4fc08973c3782634"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.291319 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-mw5qw"] Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.311877 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" event={"ID":"7d1875ce-7ea7-47ee-a495-d1d51c0b56e4","Type":"ContainerStarted","Data":"9aae2618c2cb4829c8c5c66f2b018c9bee54668665b61500727b5a6c76519398"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.320821 4959 patch_prober.go:28] interesting pod/console-operator-58897d9998-fzwkw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.320879 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" podUID="7d1875ce-7ea7-47ee-a495-d1d51c0b56e4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.323325 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.324658 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.824645631 +0000 UTC m=+146.027989048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.331486 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wlc98" podStartSLOduration=125.33145416 podStartE2EDuration="2m5.33145416s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.321342195 +0000 UTC m=+145.524685612" watchObservedRunningTime="2025-10-03 13:32:56.33145416 +0000 UTC m=+145.534797587" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.344503 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" event={"ID":"daf6334f-8900-455a-95d7-8cc149fecc58","Type":"ContainerStarted","Data":"d063e894c14e109b04e5ef0ab0e656b4ac5fef5880ca265ccea470443fc56ca0"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.359115 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" event={"ID":"289092c0-4c05-4c6d-bb0f-d36cb003b23e","Type":"ContainerStarted","Data":"b4c0dea9d528bb2399037a70e5557bbad699249469ecbdcc0060501b2251286d"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.367226 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg"] Oct 03 13:32:56 crc kubenswrapper[4959]: W1003 13:32:56.375821 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb9df3df_527a_4573_b1ac_4bf2bd94444e.slice/crio-f78b157b4a6044a1bf257dceac3ec85d3652dca1df7e1706c26e9c36f2717141 WatchSource:0}: Error finding container f78b157b4a6044a1bf257dceac3ec85d3652dca1df7e1706c26e9c36f2717141: Status 404 returned error can't find the container with id f78b157b4a6044a1bf257dceac3ec85d3652dca1df7e1706c26e9c36f2717141 Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.395276 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" event={"ID":"2cebb480-f94a-4fef-b128-281428436da2","Type":"ContainerStarted","Data":"389d1c057ad24e2a2d11dc5244f45db7d3ec3114e5109f53e4682a8aa2195e75"} Oct 03 13:32:56 crc kubenswrapper[4959]: W1003 13:32:56.395769 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod132cc225_bf51_49a1_8a9d_0058984db61e.slice/crio-b5552496a3276330612fd5317596aca615cc97f69c2720f6d9c825811bf95d56 WatchSource:0}: Error finding container b5552496a3276330612fd5317596aca615cc97f69c2720f6d9c825811bf95d56: Status 404 returned error can't find the container with id b5552496a3276330612fd5317596aca615cc97f69c2720f6d9c825811bf95d56 Oct 03 13:32:56 crc kubenswrapper[4959]: W1003 13:32:56.428721 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb04c7d45_8cef_4c51_8821_0eac0a6f6b07.slice/crio-4a8e2b1ba5d93463fef90f1c908cf6927052faf444f9e44b02a26822b362755c WatchSource:0}: Error finding container 4a8e2b1ba5d93463fef90f1c908cf6927052faf444f9e44b02a26822b362755c: Status 404 returned error can't find the container with id 4a8e2b1ba5d93463fef90f1c908cf6927052faf444f9e44b02a26822b362755c Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.429155 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.435446 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.935407633 +0000 UTC m=+146.138751050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.436154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.438990 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:56.938976977 +0000 UTC m=+146.142320394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.454522 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" event={"ID":"50e95aa6-57b0-4b4d-9d6c-c48c129c1052","Type":"ContainerStarted","Data":"1b1fae3d96cb07857f0495ef7156b60e0b641412132741e91f17fb1c99c975b1"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.454603 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" event={"ID":"50e95aa6-57b0-4b4d-9d6c-c48c129c1052","Type":"ContainerStarted","Data":"b38c26063a2968a0f5fe80c789baac36d62a6c41b20dbb9aaf48eea284c6bcd2"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.455919 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:56 crc kubenswrapper[4959]: W1003 13:32:56.479445 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e75817c_456c_4d5c_981d_c434adca9411.slice/crio-7d68caabaeb6d29dbe5c5f059ac1cbfc255c014b0dd62d4b92dcb66e2c13608a WatchSource:0}: Error finding container 7d68caabaeb6d29dbe5c5f059ac1cbfc255c014b0dd62d4b92dcb66e2c13608a: Status 404 returned error can't find the container with id 7d68caabaeb6d29dbe5c5f059ac1cbfc255c014b0dd62d4b92dcb66e2c13608a Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.487924 4959 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bcgsg container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.487922 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-xn7cr" podStartSLOduration=126.487905474 podStartE2EDuration="2m6.487905474s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.412551582 +0000 UTC m=+145.615895019" watchObservedRunningTime="2025-10-03 13:32:56.487905474 +0000 UTC m=+145.691248891" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.487963 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" podUID="50e95aa6-57b0-4b4d-9d6c-c48c129c1052" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.519051 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bg4hd" event={"ID":"704c868e-4e03-4945-9d88-382b59d8ffd1","Type":"ContainerStarted","Data":"ff54ff9cb5aaa60af4b3186aaceff6e808f4d08bf0fbb6746456b729c1a94cfe"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.521939 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" event={"ID":"5c097a19-754b-465c-8c78-fc8b9ea45526","Type":"ContainerStarted","Data":"01cf25b14304ceb23ebffed4954c03dfcc79b2336e67742d72f0b7442e6ea761"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.528904 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" event={"ID":"3f3cf4c7-106d-4f21-9596-6bd2356af306","Type":"ContainerStarted","Data":"bac04890d617c9fc5183eabcdeed877d1a407f517a977b8facf4b3e38c5b7174"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.529667 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.532203 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" event={"ID":"bac8b263-5a8c-4621-a046-496c25f91286","Type":"ContainerStarted","Data":"3cf75ffbdb12041649b97564846f9a3728ef8fa45d6294e352d33a45566e6707"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.534853 4959 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-s9jrh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.534946 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" podUID="3f3cf4c7-106d-4f21-9596-6bd2356af306" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.538255 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.539159 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.039094219 +0000 UTC m=+146.242437656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.548859 4959 generic.go:334] "Generic (PLEG): container finished" podID="f7190415-01a1-416c-b7a8-7945d9f20796" containerID="77c57e29239a2193199c2ba5c6c771dbedfe32e57035c63e2f8e3c60c1ebe1b6" exitCode=0 Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.548986 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" event={"ID":"f7190415-01a1-416c-b7a8-7945d9f20796","Type":"ContainerDied","Data":"77c57e29239a2193199c2ba5c6c771dbedfe32e57035c63e2f8e3c60c1ebe1b6"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.550906 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" event={"ID":"0061d337-ce1d-4925-b9d4-c6805aafe7ec","Type":"ContainerStarted","Data":"b242e08531d274fc40fc1289bca88421d0064f39323068d77ebad2a092ee5f54"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.554483 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bg4hd" podStartSLOduration=6.554465943 podStartE2EDuration="6.554465943s" podCreationTimestamp="2025-10-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.554406401 +0000 UTC m=+145.757749818" watchObservedRunningTime="2025-10-03 13:32:56.554465943 +0000 UTC m=+145.757809360" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.557968 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" podStartSLOduration=125.557950265 podStartE2EDuration="2m5.557950265s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.48817138 +0000 UTC m=+145.691514797" watchObservedRunningTime="2025-10-03 13:32:56.557950265 +0000 UTC m=+145.761293682" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.561983 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ff9j6" event={"ID":"3b5aee17-20ec-4adb-ad67-a239878063a6","Type":"ContainerStarted","Data":"a651fa62ca4241bee5f25c20940d7cbd4e59bb6a30e35e9e9a6349429f07d14d"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.563408 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" event={"ID":"df758465-1b0a-4ecb-b70c-c256357e4ec0","Type":"ContainerStarted","Data":"973b29810ffb8cbd036c7e26f2fb9caa6acb99f04acf1e90dfba480495ec75ff"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.564014 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.566016 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" event={"ID":"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0","Type":"ContainerStarted","Data":"b8f330f209b9ce8f39f0136ae2535f0f8bcb5d8d7bcc69b7f5e6924606ec4cc2"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.591283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" event={"ID":"b8f35bb3-a509-49f1-a945-0a8170569c1d","Type":"ContainerStarted","Data":"019bc81e241d3c78670f14b1053e7a7ec038afdbec1040bbcf5b6b6d216ce04b"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.591335 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" event={"ID":"b8f35bb3-a509-49f1-a945-0a8170569c1d","Type":"ContainerStarted","Data":"b90bbe2ccb4b4a587419faf592c3334f2074e89550315dede5b5ffdb12c9bda8"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.627332 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ttm4r" event={"ID":"847fd56c-ec05-4627-bd43-49add0bdc748","Type":"ContainerStarted","Data":"0e01ebabcde0d1e387e301ddd4a6c8139ceaf4136bc5be3a3612b00cdb4913e9"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.628540 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.629745 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.629791 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.640556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.641573 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" podStartSLOduration=125.641547062 podStartE2EDuration="2m5.641547062s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.629506396 +0000 UTC m=+145.832849823" watchObservedRunningTime="2025-10-03 13:32:56.641547062 +0000 UTC m=+145.844890479" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.644076 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" event={"ID":"275d0dcc-f61b-4ffe-a19e-bc3a2e996861","Type":"ContainerStarted","Data":"8498aaebabedb9ae41564579367e227a5bed66dfb022d58805665066e0e46a0e"} Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.649675 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.149648305 +0000 UTC m=+146.352991722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.663075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" event={"ID":"a398dbb8-67e9-4390-9a50-9c5dd2c62303","Type":"ContainerStarted","Data":"499455678f3b0effda031a4b784b70c52c56a77295483ae720419e14c870825f"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.670259 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" podStartSLOduration=125.670244767 podStartE2EDuration="2m5.670244767s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.669484036 +0000 UTC m=+145.872827453" watchObservedRunningTime="2025-10-03 13:32:56.670244767 +0000 UTC m=+145.873588184" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.673157 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" event={"ID":"d24f1f11-7f2c-4917-8648-cf1cc03d324b","Type":"ContainerStarted","Data":"08a022655da0a73005fb51454e679e76d674ec05266f21137c9acd95f24bbe27"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.717982 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" event={"ID":"677b5491-6c28-4625-854b-881e25662a57","Type":"ContainerStarted","Data":"2101c7d01d7ca6c9b9eb72e72ca6f81f3d744173727146c825a8bfab6f2190a4"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.718048 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" event={"ID":"677b5491-6c28-4625-854b-881e25662a57","Type":"ContainerStarted","Data":"6c84ea6d7aca584f6a37472bc0b8fc3efeb13b5402aa883e7dc1caf62cbcf68f"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.723160 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" event={"ID":"c6c97e09-88e1-4f2b-a476-19796c3d83f2","Type":"ContainerStarted","Data":"c03f2c46ded42a8499ab39c65a9bb238bb8bcbfadbee63009cd65bf723d0ce01"} Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.741837 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.761646 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.261627089 +0000 UTC m=+146.464970506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.762371 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ttm4r" podStartSLOduration=125.762349309 podStartE2EDuration="2m5.762349309s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.761807934 +0000 UTC m=+145.965151351" watchObservedRunningTime="2025-10-03 13:32:56.762349309 +0000 UTC m=+145.965692726" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.764024 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" podStartSLOduration=125.764009922 podStartE2EDuration="2m5.764009922s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.717279874 +0000 UTC m=+145.920623301" watchObservedRunningTime="2025-10-03 13:32:56.764009922 +0000 UTC m=+145.967353349" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.820514 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" podStartSLOduration=125.820492767 podStartE2EDuration="2m5.820492767s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.818918805 +0000 UTC m=+146.022262222" watchObservedRunningTime="2025-10-03 13:32:56.820492767 +0000 UTC m=+146.023836184" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.871923 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.923668 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.423638949 +0000 UTC m=+146.626982366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.932496 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:32:56 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:32:56 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:32:56 crc kubenswrapper[4959]: healthz check failed Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.932566 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.951321 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" podStartSLOduration=125.951300706 podStartE2EDuration="2m5.951300706s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.864657268 +0000 UTC m=+146.068000685" watchObservedRunningTime="2025-10-03 13:32:56.951300706 +0000 UTC m=+146.154644123" Oct 03 13:32:56 crc kubenswrapper[4959]: I1003 13:32:56.988734 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:56 crc kubenswrapper[4959]: E1003 13:32:56.989456 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.489436278 +0000 UTC m=+146.692779695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.023581 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" podStartSLOduration=126.023563345 podStartE2EDuration="2m6.023563345s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:56.951159052 +0000 UTC m=+146.154502469" watchObservedRunningTime="2025-10-03 13:32:57.023563345 +0000 UTC m=+146.226906762" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.037715 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" podStartSLOduration=126.037695677 podStartE2EDuration="2m6.037695677s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:57.026973215 +0000 UTC m=+146.230316632" watchObservedRunningTime="2025-10-03 13:32:57.037695677 +0000 UTC m=+146.241039094" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.089421 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" podStartSLOduration=127.089402936 podStartE2EDuration="2m7.089402936s" podCreationTimestamp="2025-10-03 13:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:57.088893253 +0000 UTC m=+146.292236670" watchObservedRunningTime="2025-10-03 13:32:57.089402936 +0000 UTC m=+146.292746353" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.090346 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.090646 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.590634738 +0000 UTC m=+146.793978155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.197833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.198174 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.698154445 +0000 UTC m=+146.901497862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.272501 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.274970 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.300772 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.301159 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.801145333 +0000 UTC m=+147.004488750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.323990 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.324063 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.402326 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.402943 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:57.902900498 +0000 UTC m=+147.106243915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.505868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.506254 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.006238405 +0000 UTC m=+147.209581822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.573308 4959 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fpgj7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.573375 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" podUID="df758465-1b0a-4ecb-b70c-c256357e4ec0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.611864 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.612674 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.112655952 +0000 UTC m=+147.315999359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.714790 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.715097 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.215085415 +0000 UTC m=+147.418428832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.760313 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z2kjq" event={"ID":"d3845ba1-19b1-4a7b-996c-ca26ffd8cda0","Type":"ContainerStarted","Data":"74b649eac59d60949e93863870f7b4dc3efa9d5bb5160970bbcd3c34e844d7cc"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.780008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" event={"ID":"60628569-ded6-49cb-8bb3-b814b5afc84a","Type":"ContainerStarted","Data":"e0ffa9b8b432d24faf4c20f07dbf0ca8e69992ac355e1fc573af40c0dda0a643"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.789863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l2prt" event={"ID":"d24f1f11-7f2c-4917-8648-cf1cc03d324b","Type":"ContainerStarted","Data":"c19d20c6e70159ea2ab0666589bf2889b6fbc9f0a82c1eafa0358c2b7e2a5401"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.804983 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-452zl" event={"ID":"0d0f85d6-3dec-4c08-8087-4abab2a3ab53","Type":"ContainerStarted","Data":"48154b5db5e5b42218205b0fb73bb527d61284ffa4816d6a515e57f368548460"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.805042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-452zl" event={"ID":"0d0f85d6-3dec-4c08-8087-4abab2a3ab53","Type":"ContainerStarted","Data":"f453116a9e76a626267b5222fd694089084dce1f20eee487fd81af393ecc5a1f"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.807040 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" event={"ID":"db9df3df-527a-4573-b1ac-4bf2bd94444e","Type":"ContainerStarted","Data":"86a50daf1a17e334aaa562d6768b76a1b960227deece0c7a386ea0dff2e760fd"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.807070 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" event={"ID":"db9df3df-527a-4573-b1ac-4bf2bd94444e","Type":"ContainerStarted","Data":"f78b157b4a6044a1bf257dceac3ec85d3652dca1df7e1706c26e9c36f2717141"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.816876 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.817276 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.317257831 +0000 UTC m=+147.520601248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.848808 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h4pbx" event={"ID":"bac8b263-5a8c-4621-a046-496c25f91286","Type":"ContainerStarted","Data":"2097841bad197c9222caa9f941c5e8aa3048123d3f392d468aeecf9e1f41a69f"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.860723 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" event={"ID":"275d0dcc-f61b-4ffe-a19e-bc3a2e996861","Type":"ContainerStarted","Data":"201f3b1f8dfb573af1b61bcadb42cf5221dd7cba0312c4fe2de6943652802b61"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.868930 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" event={"ID":"0061d337-ce1d-4925-b9d4-c6805aafe7ec","Type":"ContainerStarted","Data":"bc979d0e4194eff76923c69a30f99f7d28167c37f8714b7dbf874c4325c41d41"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.904015 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:32:57 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:32:57 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:32:57 crc kubenswrapper[4959]: healthz check failed Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.904432 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.904897 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cg2fr" event={"ID":"b8f35bb3-a509-49f1-a945-0a8170569c1d","Type":"ContainerStarted","Data":"38662992ff97ed656047ceff7c5d2fa4d28c9f5c8536230b47e98228c61d9076"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.927281 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.928315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mw5qw" event={"ID":"132cc225-bf51-49a1-8a9d-0058984db61e","Type":"ContainerStarted","Data":"b5552496a3276330612fd5317596aca615cc97f69c2720f6d9c825811bf95d56"} Oct 03 13:32:57 crc kubenswrapper[4959]: E1003 13:32:57.928548 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.428537167 +0000 UTC m=+147.631880584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.964524 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" event={"ID":"b04c7d45-8cef-4c51-8821-0eac0a6f6b07","Type":"ContainerStarted","Data":"08bedab1c821f4887836ec2c7a4fcc64988760dcc9c076a80d7c8bd4d34ad582"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.964593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" event={"ID":"b04c7d45-8cef-4c51-8821-0eac0a6f6b07","Type":"ContainerStarted","Data":"4a8e2b1ba5d93463fef90f1c908cf6927052faf444f9e44b02a26822b362755c"} Oct 03 13:32:57 crc kubenswrapper[4959]: I1003 13:32:57.977582 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-452zl" podStartSLOduration=7.977559396 podStartE2EDuration="7.977559396s" podCreationTimestamp="2025-10-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:57.976293462 +0000 UTC m=+147.179636879" watchObservedRunningTime="2025-10-03 13:32:57.977559396 +0000 UTC m=+147.180902813" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.002845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" event={"ID":"c6c97e09-88e1-4f2b-a476-19796c3d83f2","Type":"ContainerStarted","Data":"929bde24737df2c8c759f0b00eb1fbc144bfc33e1594131a41305ab4ac1908fd"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.031739 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.033229 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.533185177 +0000 UTC m=+147.736528594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.036085 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" event={"ID":"d8713790-50b4-4939-b47d-27927aefcddc","Type":"ContainerStarted","Data":"a43cf8ce9f273837c4077f34b5d050ff93fc048d9b8226694dd43c05d99fc575"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.036135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" event={"ID":"d8713790-50b4-4939-b47d-27927aefcddc","Type":"ContainerStarted","Data":"d599444c85f1ea0ea59c0b2873e643f0d0be3b503fdddff287add82ea1b0ee2b"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.036696 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.041457 4959 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gspcs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.041518 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.070495 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" event={"ID":"f7190415-01a1-416c-b7a8-7945d9f20796","Type":"ContainerStarted","Data":"17e352f9d54407acc41d9cdc4b2cbd3c49667ec9aebb5c950c4971fbbd4d16a5"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.071279 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.073215 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" event={"ID":"2cebb480-f94a-4fef-b128-281428436da2","Type":"ContainerStarted","Data":"d4b9435ab30dc1224e4dc3a38642efbabdab898651c45c13246a082ed5467295"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.091946 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zsxvg" podStartSLOduration=127.091929372 podStartE2EDuration="2m7.091929372s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.089994101 +0000 UTC m=+147.293337528" watchObservedRunningTime="2025-10-03 13:32:58.091929372 +0000 UTC m=+147.295272789" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.094182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" event={"ID":"a398dbb8-67e9-4390-9a50-9c5dd2c62303","Type":"ContainerStarted","Data":"cd341ce7e7302fe479bb708a203c2bd3aa66da9da657d87013a877fcd7263ec8"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.094249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" event={"ID":"a398dbb8-67e9-4390-9a50-9c5dd2c62303","Type":"ContainerStarted","Data":"88bf0d61bd57ee20b706ec1098bb7bc4b70111b5be89407957b78ff21f29dbb3"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.094965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.096769 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" event={"ID":"289092c0-4c05-4c6d-bb0f-d36cb003b23e","Type":"ContainerStarted","Data":"dac47b407e73edf118797741433a9871b62f8389b41488afe95492ebb03b1af2"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.096815 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" event={"ID":"289092c0-4c05-4c6d-bb0f-d36cb003b23e","Type":"ContainerStarted","Data":"9a1d57f674094771a4391ce73ff844ea5f68a652d17643b11613e631f8d9e309"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.110394 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" event={"ID":"67e98663-cc71-4af8-8a92-446851e11efe","Type":"ContainerStarted","Data":"83955c306fdd53020e5c8e4fa5e261fe4eb04d8baac5f4966154482313ab0995"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.110439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" event={"ID":"67e98663-cc71-4af8-8a92-446851e11efe","Type":"ContainerStarted","Data":"0c1ae59ed63dee3b56f5dc227c68b3bfd141c7f26c8819ced7b91f4a33501fb3"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.117395 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" event={"ID":"2e75817c-456c-4d5c-981d-c434adca9411","Type":"ContainerStarted","Data":"c647f354b4304b57073c7831567496c29f3f89678d2d6dc3cc4acdc3d6a57923"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.117443 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" event={"ID":"2e75817c-456c-4d5c-981d-c434adca9411","Type":"ContainerStarted","Data":"7d68caabaeb6d29dbe5c5f059ac1cbfc255c014b0dd62d4b92dcb66e2c13608a"} Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.126695 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.126749 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.133381 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.134790 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.634778579 +0000 UTC m=+147.838121996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.140441 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.189102 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s9jrh" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.191436 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fpgj7" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.232826 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bcgsg" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.237838 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.239826 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.73980372 +0000 UTC m=+147.943147147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.312517 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f74nt" podStartSLOduration=127.3124906 podStartE2EDuration="2m7.3124906s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.191245943 +0000 UTC m=+147.394589360" watchObservedRunningTime="2025-10-03 13:32:58.3124906 +0000 UTC m=+147.515834017" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.334888 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" podStartSLOduration=127.334868369 podStartE2EDuration="2m7.334868369s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.325514343 +0000 UTC m=+147.528857760" watchObservedRunningTime="2025-10-03 13:32:58.334868369 +0000 UTC m=+147.538211786" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.343626 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.344218 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.844178794 +0000 UTC m=+148.047522211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.351810 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9l2d2" podStartSLOduration=127.351785523 podStartE2EDuration="2m7.351785523s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.294565309 +0000 UTC m=+147.497908726" watchObservedRunningTime="2025-10-03 13:32:58.351785523 +0000 UTC m=+147.555128940" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.369204 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" podStartSLOduration=127.36917374 podStartE2EDuration="2m7.36917374s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.36840658 +0000 UTC m=+147.571749997" watchObservedRunningTime="2025-10-03 13:32:58.36917374 +0000 UTC m=+147.572517157" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.444762 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.445232 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:58.94520969 +0000 UTC m=+148.148553107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.497212 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" podStartSLOduration=127.497178686 podStartE2EDuration="2m7.497178686s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.480657661 +0000 UTC m=+147.684001078" watchObservedRunningTime="2025-10-03 13:32:58.497178686 +0000 UTC m=+147.700522103" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.532316 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.533238 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.538909 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.546391 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.546672 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.046661056 +0000 UTC m=+148.250004473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.558726 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-thbnm" podStartSLOduration=127.558708513 podStartE2EDuration="2m7.558708513s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.547307564 +0000 UTC m=+147.750651001" watchObservedRunningTime="2025-10-03 13:32:58.558708513 +0000 UTC m=+147.762051920" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.560534 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-fzwkw" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.561025 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.649737 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.650300 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.650352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.650398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6jrn\" (UniqueName: \"kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.650610 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.150587388 +0000 UTC m=+148.353930815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.662524 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-47mhg" podStartSLOduration=127.662502462 podStartE2EDuration="2m7.662502462s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.609450667 +0000 UTC m=+147.812794084" watchObservedRunningTime="2025-10-03 13:32:58.662502462 +0000 UTC m=+147.865845889" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.707636 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" podStartSLOduration=127.707619848 podStartE2EDuration="2m7.707619848s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.704958368 +0000 UTC m=+147.908301775" watchObservedRunningTime="2025-10-03 13:32:58.707619848 +0000 UTC m=+147.910963265" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.720298 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.721105 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.724446 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752133 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752183 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752218 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752238 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752276 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.752303 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6jrn\" (UniqueName: \"kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.757056 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.758586 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.759817 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.259800919 +0000 UTC m=+148.463144336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.761406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.761446 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.763438 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.775311 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" podStartSLOduration=127.775289827 podStartE2EDuration="2m7.775289827s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.741400166 +0000 UTC m=+147.944743593" watchObservedRunningTime="2025-10-03 13:32:58.775289827 +0000 UTC m=+147.978633244" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.775982 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.778132 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.815790 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6jrn\" (UniqueName: \"kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn\") pod \"certified-operators-49wzf\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.854049 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.854343 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.854407 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.854511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjrsz\" (UniqueName: \"kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.854665 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.354637393 +0000 UTC m=+148.557980810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.867465 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.895395 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" podStartSLOduration=127.895371424 podStartE2EDuration="2m7.895371424s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.834943175 +0000 UTC m=+148.038286602" watchObservedRunningTime="2025-10-03 13:32:58.895371424 +0000 UTC m=+148.098714841" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.908510 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:32:58 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:32:58 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:32:58 crc kubenswrapper[4959]: healthz check failed Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.908564 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.909140 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.910499 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vms5l" podStartSLOduration=127.910478421 podStartE2EDuration="2m7.910478421s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:58.896975816 +0000 UTC m=+148.100319233" watchObservedRunningTime="2025-10-03 13:32:58.910478421 +0000 UTC m=+148.113821838" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.913530 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.914743 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.937918 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.957124 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjrsz\" (UniqueName: \"kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.957202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.957228 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.957274 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:58 crc kubenswrapper[4959]: E1003 13:32:58.957617 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.45760173 +0000 UTC m=+148.660945157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.958358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.958632 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:58 crc kubenswrapper[4959]: I1003 13:32:58.995452 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjrsz\" (UniqueName: \"kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz\") pod \"community-operators-lhm6k\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.009931 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.025261 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.056561 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.061929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.062141 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.062208 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knxkc\" (UniqueName: \"kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.062311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.062447 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.562425195 +0000 UTC m=+148.765768612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.112874 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.123749 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.123875 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.164639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.164690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knxkc\" (UniqueName: \"kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.164926 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.164962 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.169136 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.66911541 +0000 UTC m=+148.872458837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.169796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.171163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.191524 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-v6d5j" event={"ID":"2cebb480-f94a-4fef-b128-281428436da2","Type":"ContainerStarted","Data":"764f8b3abae0a0d2a7e744f3c010bbdd1761754e3e183995a1864805fb24a7ce"} Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.201971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knxkc\" (UniqueName: \"kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc\") pod \"certified-operators-jnggp\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.245990 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mw5qw" event={"ID":"132cc225-bf51-49a1-8a9d-0058984db61e","Type":"ContainerStarted","Data":"c624192f81cc12af0f350505c396525fa9055c0983f4bd103afc9cd8572987f7"} Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.246043 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-mw5qw" event={"ID":"132cc225-bf51-49a1-8a9d-0058984db61e","Type":"ContainerStarted","Data":"478f97c15d6bfb08b7a8c36b708d07ae035da785db0fd14777f0a746b4236680"} Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.246857 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-mw5qw" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.259673 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" event={"ID":"60628569-ded6-49cb-8bb3-b814b5afc84a","Type":"ContainerStarted","Data":"08e231c1154b6a2662a7feae297a777662fb3ff4c6500ec0a1a05940993803e5"} Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.268802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.269038 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.269137 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.269253 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ngvn\" (UniqueName: \"kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.269386 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.769363405 +0000 UTC m=+148.972706822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.271332 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j5tgf" event={"ID":"b04c7d45-8cef-4c51-8821-0eac0a6f6b07","Type":"ContainerStarted","Data":"2675d9b750b63f325cd2724f63e34427934c3f7bca2ba56ec994c8b8395646b4"} Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.272736 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.272779 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.274668 4959 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gspcs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.274706 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.278526 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.288066 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-mw5qw" podStartSLOduration=9.288048577 podStartE2EDuration="9.288048577s" podCreationTimestamp="2025-10-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:32:59.286588689 +0000 UTC m=+148.489932126" watchObservedRunningTime="2025-10-03 13:32:59.288048577 +0000 UTC m=+148.491391994" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.327392 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5l9tb" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.375848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.376357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.376497 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ngvn\" (UniqueName: \"kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.376637 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.381153 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.389456 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.889442152 +0000 UTC m=+149.092785559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.392513 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.478836 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.479454 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:32:59.979431338 +0000 UTC m=+149.182774755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.489915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ngvn\" (UniqueName: \"kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn\") pod \"community-operators-kmtqr\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.509163 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.580803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.581227 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.081209374 +0000 UTC m=+149.284552791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.683699 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.684006 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.183967495 +0000 UTC m=+149.387310922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.684308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.684615 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.184606942 +0000 UTC m=+149.387950359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.758287 4959 patch_prober.go:28] interesting pod/apiserver-76f77b778f-h7s26 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]log ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]etcd ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/generic-apiserver-start-informers ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/max-in-flight-filter ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 03 13:32:59 crc kubenswrapper[4959]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 03 13:32:59 crc kubenswrapper[4959]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/project.openshift.io-projectcache ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 03 13:32:59 crc kubenswrapper[4959]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 03 13:32:59 crc kubenswrapper[4959]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 03 13:32:59 crc kubenswrapper[4959]: livez check failed Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.758712 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" podUID="677b5491-6c28-4625-854b-881e25662a57" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.790024 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.790341 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.290324681 +0000 UTC m=+149.493668098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.892864 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:32:59 crc kubenswrapper[4959]: E1003 13:32:59.893230 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.393214666 +0000 UTC m=+149.596558093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.902452 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:32:59 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:32:59 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:32:59 crc kubenswrapper[4959]: healthz check failed Oct 03 13:32:59 crc kubenswrapper[4959]: I1003 13:32:59.902519 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.000007 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.000328 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.500312372 +0000 UTC m=+149.703655789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.010649 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:33:00 crc kubenswrapper[4959]: W1003 13:33:00.047385 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78e585b_e57b_4681_863c_6489cc654222.slice/crio-2fe743f297112f8470905b13cc1d04885f88857862619de4cf7bc20b95956444 WatchSource:0}: Error finding container 2fe743f297112f8470905b13cc1d04885f88857862619de4cf7bc20b95956444: Status 404 returned error can't find the container with id 2fe743f297112f8470905b13cc1d04885f88857862619de4cf7bc20b95956444 Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.104860 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.105203 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.605177469 +0000 UTC m=+149.808520886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.206126 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.206891 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.706849192 +0000 UTC m=+149.910192609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.278340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"eea022fd4506fad2a16a71d1d551a868da2fd289093688669421e2c52d4b9234"} Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.298583 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.308109 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.308552 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.808535864 +0000 UTC m=+150.011879281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.322469 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerStarted","Data":"2fe743f297112f8470905b13cc1d04885f88857862619de4cf7bc20b95956444"} Oct 03 13:33:00 crc kubenswrapper[4959]: W1003 13:33:00.338263 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod588d8af1_ff2a_44de_ad5f_2fb64b934cef.slice/crio-3041b63fcd4dafb961171958912d72d281206f73b90651fbfb4c686d95d284fe WatchSource:0}: Error finding container 3041b63fcd4dafb961171958912d72d281206f73b90651fbfb4c686d95d284fe: Status 404 returned error can't find the container with id 3041b63fcd4dafb961171958912d72d281206f73b90651fbfb4c686d95d284fe Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.409330 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.411751 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:00.911729688 +0000 UTC m=+150.115073105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.450986 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.515968 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.516326 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.016311327 +0000 UTC m=+150.219654744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.517253 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.518118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.530050 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.532680 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.568894 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.616934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.617085 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.117059896 +0000 UTC m=+150.320403313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.617288 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.617344 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwgpk\" (UniqueName: \"kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.617364 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.617417 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.617657 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.117646081 +0000 UTC m=+150.320989508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: W1003 13:33:00.715527 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60a8d300_9415_476c_871b_ef8400c8656c.slice/crio-dacc5d77bac287313afa9b5fdaface198b330e5e3b284769b768bb536fdadfea WatchSource:0}: Error finding container dacc5d77bac287313afa9b5fdaface198b330e5e3b284769b768bb536fdadfea: Status 404 returned error can't find the container with id dacc5d77bac287313afa9b5fdaface198b330e5e3b284769b768bb536fdadfea Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.717988 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.718141 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.218115362 +0000 UTC m=+150.421458789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718281 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwgpk\" (UniqueName: \"kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718644 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.718676 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.718710 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.218690257 +0000 UTC m=+150.422033674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.751265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwgpk\" (UniqueName: \"kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk\") pod \"redhat-marketplace-7zsrk\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.767974 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-nqgtl" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.819730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.820453 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.320432022 +0000 UTC m=+150.523775439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.859711 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.901724 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:00 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:00 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:00 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.901782 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.916627 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.917905 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.922980 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:00 crc kubenswrapper[4959]: E1003 13:33:00.923393 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.423378119 +0000 UTC m=+150.626721536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:00 crc kubenswrapper[4959]: I1003 13:33:00.930528 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.023659 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.024085 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.024148 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.024225 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gn92\" (UniqueName: \"kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.024384 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.524345383 +0000 UTC m=+150.727688800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.062181 4959 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125011 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125074 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125091 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gn92\" (UniqueName: \"kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125150 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.125755 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.126055 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.626042396 +0000 UTC m=+150.829385813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.153025 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gn92\" (UniqueName: \"kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92\") pod \"redhat-marketplace-jzzcf\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.168258 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.226442 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.226937 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.726913478 +0000 UTC m=+150.930256895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: W1003 13:33:01.244339 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bee80ae_896e_4d7d_885d_e1bf8280db86.slice/crio-40a1d1e61ce721e87d4c5050ef2c01a9abd29f82f7601e1d7a7abdc83a706fb4 WatchSource:0}: Error finding container 40a1d1e61ce721e87d4c5050ef2c01a9abd29f82f7601e1d7a7abdc83a706fb4: Status 404 returned error can't find the container with id 40a1d1e61ce721e87d4c5050ef2c01a9abd29f82f7601e1d7a7abdc83a706fb4 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.307509 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.328922 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.328964 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerStarted","Data":"40a1d1e61ce721e87d4c5050ef2c01a9abd29f82f7601e1d7a7abdc83a706fb4"} Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.329298 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.82928469 +0000 UTC m=+151.032628107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.330604 4959 generic.go:334] "Generic (PLEG): container finished" podID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerID="d56705f092f97d78c98817fb48e3cacabedde398f7c7f20d821ace3c14127a95" exitCode=0 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.330665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerDied","Data":"d56705f092f97d78c98817fb48e3cacabedde398f7c7f20d821ace3c14127a95"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.330697 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerStarted","Data":"3041b63fcd4dafb961171958912d72d281206f73b90651fbfb4c686d95d284fe"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.332773 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.339082 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1731b9388202859830328e52ad2d53c2f367baa00db30812ba3cbfadfb037169"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.339295 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"31c5e8e94c1a9921354a7601738da762519e6029bba84780c9c9bb871fed9e2e"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.351119 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" event={"ID":"60628569-ded6-49cb-8bb3-b814b5afc84a","Type":"ContainerStarted","Data":"a8e9edf3cb77ecdfef654f11834f15395757532cc6689ca562c0311045ae8cf4"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.351320 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" event={"ID":"60628569-ded6-49cb-8bb3-b814b5afc84a","Type":"ContainerStarted","Data":"c5262e7e43ba1cc2bc24344821573780c89d931ac1783057ba00efb5eede5b4d"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.352779 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8819c38a09dfa672cb0ba8282bca23de88c4881fddcc939222e422a0b6366455"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.353491 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.354717 4959 generic.go:334] "Generic (PLEG): container finished" podID="f78e585b-e57b-4681-863c-6489cc654222" containerID="40a4a7dac42e52f655b25a9cfb8f33dda11b5047745922bd962c00664a2d6138" exitCode=0 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.354805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerDied","Data":"40a4a7dac42e52f655b25a9cfb8f33dda11b5047745922bd962c00664a2d6138"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.375475 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerID="1452e11e23cf35873af0fc9da5542fa42b4b26f93d3db85ec314a77664781c73" exitCode=0 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.375551 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerDied","Data":"1452e11e23cf35873af0fc9da5542fa42b4b26f93d3db85ec314a77664781c73"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.375600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerStarted","Data":"d70439f58b96bd366813c5dee4946f692ef8494016978edc1781a54c76ca8470"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.380169 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"af73f3b4d9498fcadfc1870bfd1381bffdae4c86f57279ee93d12f8762623e7a"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.380229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8d5401cc40578a0e7e47cce20bc6ba8da9ab7a8cc8da08bb0dba9ea7dfbced33"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.389096 4959 generic.go:334] "Generic (PLEG): container finished" podID="60a8d300-9415-476c-871b-ef8400c8656c" containerID="e7c69e414edc1506dd8c7681e6401e6c3e8115dc01645ab8a2829f5970e889de" exitCode=0 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.390604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerDied","Data":"e7c69e414edc1506dd8c7681e6401e6c3e8115dc01645ab8a2829f5970e889de"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.390639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerStarted","Data":"dacc5d77bac287313afa9b5fdaface198b330e5e3b284769b768bb536fdadfea"} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.439840 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.440108 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.940063112 +0000 UTC m=+151.143406529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.440883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.441552 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:01.941533771 +0000 UTC m=+151.144877198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.545961 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.546216 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.04614997 +0000 UTC m=+151.249493397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.546768 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.547104 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.047087315 +0000 UTC m=+151.250430732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.568075 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.648138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.648331 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.148303096 +0000 UTC m=+151.351646513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.648648 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.648928 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.148916802 +0000 UTC m=+151.352260219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.749213 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.749435 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.249401504 +0000 UTC m=+151.452744921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.749529 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.749842 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.249831825 +0000 UTC m=+151.453175242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.850045 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.850267 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.350240455 +0000 UTC m=+151.553583872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.850353 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:01 crc kubenswrapper[4959]: E1003 13:33:01.850725 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 13:33:02.350713867 +0000 UTC m=+151.554057284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-56clw" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.871455 4959 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-03T13:33:01.062700251Z","Handler":null,"Name":""} Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.877599 4959 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.877632 4959 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.895855 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:01 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:01 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:01 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.895919 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.902506 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.903733 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.908737 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.911614 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.952140 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 13:33:01 crc kubenswrapper[4959]: I1003 13:33:01.956725 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.054624 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.054725 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8gpv\" (UniqueName: \"kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.054919 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.055026 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.060828 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.060867 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.085491 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-56clw\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.156442 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.156561 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.156636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8gpv\" (UniqueName: \"kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.158012 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.158418 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.195339 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8gpv\" (UniqueName: \"kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv\") pod \"redhat-operators-l8jn9\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.220224 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.278123 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.286602 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h7s26" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.319614 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.328619 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.335398 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.373318 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.447138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" event={"ID":"60628569-ded6-49cb-8bb3-b814b5afc84a","Type":"ContainerStarted","Data":"f4288f710e6e4336cb2c34f77ccb8011e8b0393651510c77630cf4f3145a14de"} Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.451556 4959 generic.go:334] "Generic (PLEG): container finished" podID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerID="1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a" exitCode=0 Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.451702 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerDied","Data":"1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a"} Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.456619 4959 generic.go:334] "Generic (PLEG): container finished" podID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerID="6b92671e94ccae1be82d65e5968388ba378139544bfd4e989f3ae22558ae4deb" exitCode=0 Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.456816 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerDied","Data":"6b92671e94ccae1be82d65e5968388ba378139544bfd4e989f3ae22558ae4deb"} Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.456866 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerStarted","Data":"b19242379a1d10012c05a01e1d3bc06a8184253f9a913928fcacfb6ad39a45aa"} Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.465644 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckpcs\" (UniqueName: \"kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.465727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.465831 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.485166 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-v8x2x" podStartSLOduration=12.485144526 podStartE2EDuration="12.485144526s" podCreationTimestamp="2025-10-03 13:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:33:02.484044397 +0000 UTC m=+151.687387834" watchObservedRunningTime="2025-10-03 13:33:02.485144526 +0000 UTC m=+151.688487943" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.567222 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckpcs\" (UniqueName: \"kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.567639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.567836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.568584 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.571616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.610812 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckpcs\" (UniqueName: \"kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs\") pod \"redhat-operators-tw6vm\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.699668 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.852110 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.852747 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.857651 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.857932 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.873933 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.888897 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.896976 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.904511 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:02 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:02 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:02 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.904575 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.982623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:02 crc kubenswrapper[4959]: I1003 13:33:02.982872 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.012107 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077582 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077622 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077649 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077691 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077723 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.077715 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.080765 4959 patch_prober.go:28] interesting pod/console-f9d7485db-l2fnx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.080811 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l2fnx" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.083788 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.083868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.084210 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.135043 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.208680 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.239062 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:33:03 crc kubenswrapper[4959]: W1003 13:33:03.259395 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7a296e7_4d6f_40de_a3f3_567cc0bdedcc.slice/crio-cdc9bcec9cea9698f8816e0377a87bfacab47369b1eadc8ba3fa830fa5d177be WatchSource:0}: Error finding container cdc9bcec9cea9698f8816e0377a87bfacab47369b1eadc8ba3fa830fa5d177be: Status 404 returned error can't find the container with id cdc9bcec9cea9698f8816e0377a87bfacab47369b1eadc8ba3fa830fa5d177be Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.494466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerStarted","Data":"cdc9bcec9cea9698f8816e0377a87bfacab47369b1eadc8ba3fa830fa5d177be"} Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.499875 4959 generic.go:334] "Generic (PLEG): container finished" podID="0324bc20-8d56-4195-b566-aad202e4fe70" containerID="15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212" exitCode=0 Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.499943 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerDied","Data":"15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212"} Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.499975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerStarted","Data":"56620885646bb078d74a485bfe68c8c6fc87757377e311ba75d96fcc43080846"} Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.507429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" event={"ID":"9f722c0f-dd0f-4d39-a444-687158d69894","Type":"ContainerStarted","Data":"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7"} Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.507493 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" event={"ID":"9f722c0f-dd0f-4d39-a444-687158d69894","Type":"ContainerStarted","Data":"e70712fadc1104751d075d9dbced34d48692e028839d8584e60b21803d510be2"} Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.507591 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.531262 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.554502 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" podStartSLOduration=132.554478528 podStartE2EDuration="2m12.554478528s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:33:03.546415016 +0000 UTC m=+152.749758453" watchObservedRunningTime="2025-10-03 13:33:03.554478528 +0000 UTC m=+152.757821945" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.707235 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.796912 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.901411 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:03 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:03 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:03 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:03 crc kubenswrapper[4959]: I1003 13:33:03.901470 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:04 crc kubenswrapper[4959]: I1003 13:33:04.552024 4959 generic.go:334] "Generic (PLEG): container finished" podID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerID="d0037b2450eb5432cf9e0ea2e7c6099282aab12b9e271c818dc4724352ce1767" exitCode=0 Oct 03 13:33:04 crc kubenswrapper[4959]: I1003 13:33:04.552249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerDied","Data":"d0037b2450eb5432cf9e0ea2e7c6099282aab12b9e271c818dc4724352ce1767"} Oct 03 13:33:04 crc kubenswrapper[4959]: I1003 13:33:04.556656 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1460159d-ec5f-4b4a-821a-9c0a492b8d0d","Type":"ContainerStarted","Data":"46becb69d5189f00ad401f64b838fa55deb06eee3c86973a88ca58ef772e67ce"} Oct 03 13:33:04 crc kubenswrapper[4959]: I1003 13:33:04.896825 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:04 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:04 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:04 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:04 crc kubenswrapper[4959]: I1003 13:33:04.896895 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.048974 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.160804 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.161723 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.171625 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.171883 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.180996 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.223420 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.223475 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.329913 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.329986 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.330151 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.374939 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.488321 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.589261 4959 generic.go:334] "Generic (PLEG): container finished" podID="1460159d-ec5f-4b4a-821a-9c0a492b8d0d" containerID="91e62db70bf16f67a8d8650c73562e7edbf57e57aee54b32ef845fb55138cda6" exitCode=0 Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.589315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1460159d-ec5f-4b4a-821a-9c0a492b8d0d","Type":"ContainerDied","Data":"91e62db70bf16f67a8d8650c73562e7edbf57e57aee54b32ef845fb55138cda6"} Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.860298 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.900014 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:05 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:05 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:05 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:05 crc kubenswrapper[4959]: I1003 13:33:05.900099 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:06 crc kubenswrapper[4959]: I1003 13:33:06.044685 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:33:06 crc kubenswrapper[4959]: I1003 13:33:06.044755 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:33:06 crc kubenswrapper[4959]: I1003 13:33:06.619562 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97f92826-313a-4192-b1f3-479f751ebf20","Type":"ContainerStarted","Data":"2d3735c3da83ea22c9a186775b568d156c0bc46d833493323a504e15af012ce1"} Oct 03 13:33:06 crc kubenswrapper[4959]: I1003 13:33:06.896694 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:06 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:06 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:06 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:06 crc kubenswrapper[4959]: I1003 13:33:06.896761 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.137072 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.165710 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access\") pod \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.165849 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir\") pod \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\" (UID: \"1460159d-ec5f-4b4a-821a-9c0a492b8d0d\") " Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.166182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1460159d-ec5f-4b4a-821a-9c0a492b8d0d" (UID: "1460159d-ec5f-4b4a-821a-9c0a492b8d0d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.175482 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1460159d-ec5f-4b4a-821a-9c0a492b8d0d" (UID: "1460159d-ec5f-4b4a-821a-9c0a492b8d0d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.267651 4959 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.267684 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1460159d-ec5f-4b4a-821a-9c0a492b8d0d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.643703 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.645053 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1460159d-ec5f-4b4a-821a-9c0a492b8d0d","Type":"ContainerDied","Data":"46becb69d5189f00ad401f64b838fa55deb06eee3c86973a88ca58ef772e67ce"} Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.645102 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46becb69d5189f00ad401f64b838fa55deb06eee3c86973a88ca58ef772e67ce" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.659551 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97f92826-313a-4192-b1f3-479f751ebf20","Type":"ContainerStarted","Data":"3ffc08e60d4997ffedbe09b3d4f69def65e46172ceff5d066989c35e7350ed1f"} Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.686883 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.686855474 podStartE2EDuration="2.686855474s" podCreationTimestamp="2025-10-03 13:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:33:07.679436729 +0000 UTC m=+156.882780156" watchObservedRunningTime="2025-10-03 13:33:07.686855474 +0000 UTC m=+156.890198891" Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.897618 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:07 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:07 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:07 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:07 crc kubenswrapper[4959]: I1003 13:33:07.897691 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.670608 4959 generic.go:334] "Generic (PLEG): container finished" podID="c6c97e09-88e1-4f2b-a476-19796c3d83f2" containerID="929bde24737df2c8c759f0b00eb1fbc144bfc33e1594131a41305ab4ac1908fd" exitCode=0 Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.670670 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" event={"ID":"c6c97e09-88e1-4f2b-a476-19796c3d83f2","Type":"ContainerDied","Data":"929bde24737df2c8c759f0b00eb1fbc144bfc33e1594131a41305ab4ac1908fd"} Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.673037 4959 generic.go:334] "Generic (PLEG): container finished" podID="97f92826-313a-4192-b1f3-479f751ebf20" containerID="3ffc08e60d4997ffedbe09b3d4f69def65e46172ceff5d066989c35e7350ed1f" exitCode=0 Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.673065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97f92826-313a-4192-b1f3-479f751ebf20","Type":"ContainerDied","Data":"3ffc08e60d4997ffedbe09b3d4f69def65e46172ceff5d066989c35e7350ed1f"} Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.684295 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-mw5qw" Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.895260 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:08 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:08 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:08 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:08 crc kubenswrapper[4959]: I1003 13:33:08.895309 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:09 crc kubenswrapper[4959]: I1003 13:33:09.897162 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:09 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:09 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:09 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:09 crc kubenswrapper[4959]: I1003 13:33:09.897452 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:10 crc kubenswrapper[4959]: I1003 13:33:10.895634 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:10 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:10 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:10 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:10 crc kubenswrapper[4959]: I1003 13:33:10.896053 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:11 crc kubenswrapper[4959]: I1003 13:33:11.894811 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:11 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:11 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:11 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:11 crc kubenswrapper[4959]: I1003 13:33:11.894889 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:12 crc kubenswrapper[4959]: I1003 13:33:12.895497 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:12 crc kubenswrapper[4959]: [-]has-synced failed: reason withheld Oct 03 13:33:12 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:12 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:12 crc kubenswrapper[4959]: I1003 13:33:12.895980 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.071023 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.071085 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.071304 4959 patch_prober.go:28] interesting pod/downloads-7954f5f757-ttm4r container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.071383 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ttm4r" podUID="847fd56c-ec05-4627-bd43-49add0bdc748" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.078169 4959 patch_prober.go:28] interesting pod/console-f9d7485db-l2fnx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.078239 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l2fnx" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.166279 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.198389 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3a679d1-7412-4843-998e-8d1c109bf7ec-metrics-certs\") pod \"network-metrics-daemon-gf4tr\" (UID: \"e3a679d1-7412-4843-998e-8d1c109bf7ec\") " pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.437683 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gf4tr" Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.894865 4959 patch_prober.go:28] interesting pod/router-default-5444994796-k89zn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 13:33:13 crc kubenswrapper[4959]: [+]has-synced ok Oct 03 13:33:13 crc kubenswrapper[4959]: [+]process-running ok Oct 03 13:33:13 crc kubenswrapper[4959]: healthz check failed Oct 03 13:33:13 crc kubenswrapper[4959]: I1003 13:33:13.895241 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-k89zn" podUID="3393e55a-a44e-42d8-a1ad-7c62ead0edc2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:33:14 crc kubenswrapper[4959]: I1003 13:33:14.895231 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:33:14 crc kubenswrapper[4959]: I1003 13:33:14.897860 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-k89zn" Oct 03 13:33:19 crc kubenswrapper[4959]: I1003 13:33:19.993341 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:33:19 crc kubenswrapper[4959]: I1003 13:33:19.999614 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158625 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir\") pod \"97f92826-313a-4192-b1f3-479f751ebf20\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158703 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwfx9\" (UniqueName: \"kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9\") pod \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158761 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") pod \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158785 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access\") pod \"97f92826-313a-4192-b1f3-479f751ebf20\" (UID: \"97f92826-313a-4192-b1f3-479f751ebf20\") " Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume\") pod \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\" (UID: \"c6c97e09-88e1-4f2b-a476-19796c3d83f2\") " Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.158801 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "97f92826-313a-4192-b1f3-479f751ebf20" (UID: "97f92826-313a-4192-b1f3-479f751ebf20"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.159048 4959 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/97f92826-313a-4192-b1f3-479f751ebf20-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.159672 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "c6c97e09-88e1-4f2b-a476-19796c3d83f2" (UID: "c6c97e09-88e1-4f2b-a476-19796c3d83f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.164092 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "97f92826-313a-4192-b1f3-479f751ebf20" (UID: "97f92826-313a-4192-b1f3-479f751ebf20"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.164370 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9" (OuterVolumeSpecName: "kube-api-access-rwfx9") pod "c6c97e09-88e1-4f2b-a476-19796c3d83f2" (UID: "c6c97e09-88e1-4f2b-a476-19796c3d83f2"). InnerVolumeSpecName "kube-api-access-rwfx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.164482 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c6c97e09-88e1-4f2b-a476-19796c3d83f2" (UID: "c6c97e09-88e1-4f2b-a476-19796c3d83f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.260068 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwfx9\" (UniqueName: \"kubernetes.io/projected/c6c97e09-88e1-4f2b-a476-19796c3d83f2-kube-api-access-rwfx9\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.260115 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6c97e09-88e1-4f2b-a476-19796c3d83f2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.260129 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/97f92826-313a-4192-b1f3-479f751ebf20-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.260141 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6c97e09-88e1-4f2b-a476-19796c3d83f2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.759780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"97f92826-313a-4192-b1f3-479f751ebf20","Type":"ContainerDied","Data":"2d3735c3da83ea22c9a186775b568d156c0bc46d833493323a504e15af012ce1"} Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.759944 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d3735c3da83ea22c9a186775b568d156c0bc46d833493323a504e15af012ce1" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.759802 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.765212 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.765110 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd" event={"ID":"c6c97e09-88e1-4f2b-a476-19796c3d83f2","Type":"ContainerDied","Data":"c03f2c46ded42a8499ab39c65a9bb238bb8bcbfadbee63009cd65bf723d0ce01"} Oct 03 13:33:20 crc kubenswrapper[4959]: I1003 13:33:20.766369 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c03f2c46ded42a8499ab39c65a9bb238bb8bcbfadbee63009cd65bf723d0ce01" Oct 03 13:33:22 crc kubenswrapper[4959]: I1003 13:33:22.333733 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:33:23 crc kubenswrapper[4959]: I1003 13:33:23.076409 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ttm4r" Oct 03 13:33:23 crc kubenswrapper[4959]: I1003 13:33:23.082830 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:33:23 crc kubenswrapper[4959]: I1003 13:33:23.090643 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:33:33 crc kubenswrapper[4959]: I1003 13:33:33.268574 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9h4nn" Oct 03 13:33:36 crc kubenswrapper[4959]: I1003 13:33:36.045029 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:33:36 crc kubenswrapper[4959]: I1003 13:33:36.045740 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:33:39 crc kubenswrapper[4959]: I1003 13:33:39.042287 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.298068 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.298824 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-knxkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jnggp_openshift-marketplace(e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.300051 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jnggp" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.878309 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.878729 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwgpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7zsrk_openshift-marketplace(1bee80ae-896e-4d7d-885d-e1bf8280db86): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:45 crc kubenswrapper[4959]: E1003 13:33:45.880012 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7zsrk" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" Oct 03 13:33:46 crc kubenswrapper[4959]: E1003 13:33:46.969472 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7zsrk" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" Oct 03 13:33:46 crc kubenswrapper[4959]: E1003 13:33:46.969543 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jnggp" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" Oct 03 13:33:47 crc kubenswrapper[4959]: E1003 13:33:47.048155 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 13:33:47 crc kubenswrapper[4959]: E1003 13:33:47.048405 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ngvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kmtqr_openshift-marketplace(60a8d300-9415-476c-871b-ef8400c8656c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:47 crc kubenswrapper[4959]: E1003 13:33:47.049570 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kmtqr" podUID="60a8d300-9415-476c-871b-ef8400c8656c" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.427503 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kmtqr" podUID="60a8d300-9415-476c-871b-ef8400c8656c" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.504451 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.504775 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ckpcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tw6vm_openshift-marketplace(c7a296e7-4d6f-40de-a3f3-567cc0bdedcc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.506184 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tw6vm" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.531687 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.531826 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gn92,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jzzcf_openshift-marketplace(8132bb0b-e372-4e00-93dd-8be8982cad6f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.533222 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jzzcf" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.534058 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.534145 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b6jrn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-49wzf_openshift-marketplace(f78e585b-e57b-4681-863c-6489cc654222): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.535272 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-49wzf" podUID="f78e585b-e57b-4681-863c-6489cc654222" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.550703 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.550921 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w8gpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-l8jn9_openshift-marketplace(0324bc20-8d56-4195-b566-aad202e4fe70): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.552156 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-l8jn9" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.558944 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.559087 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mjrsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lhm6k_openshift-marketplace(588d8af1-ff2a-44de-ad5f-2fb64b934cef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.560315 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lhm6k" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" Oct 03 13:33:49 crc kubenswrapper[4959]: I1003 13:33:49.645568 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gf4tr"] Oct 03 13:33:49 crc kubenswrapper[4959]: W1003 13:33:49.652899 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3a679d1_7412_4843_998e_8d1c109bf7ec.slice/crio-b22628da20d51599107079490319f75ca4eca7d5cf0288460a280468bb3cd2bf WatchSource:0}: Error finding container b22628da20d51599107079490319f75ca4eca7d5cf0288460a280468bb3cd2bf: Status 404 returned error can't find the container with id b22628da20d51599107079490319f75ca4eca7d5cf0288460a280468bb3cd2bf Oct 03 13:33:49 crc kubenswrapper[4959]: I1003 13:33:49.960965 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" event={"ID":"e3a679d1-7412-4843-998e-8d1c109bf7ec","Type":"ContainerStarted","Data":"d595f163a531289374a54e3a985ec621f28910dbf8c12f044e69f1258e7ac2df"} Oct 03 13:33:49 crc kubenswrapper[4959]: I1003 13:33:49.961843 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" event={"ID":"e3a679d1-7412-4843-998e-8d1c109bf7ec","Type":"ContainerStarted","Data":"b22628da20d51599107079490319f75ca4eca7d5cf0288460a280468bb3cd2bf"} Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.964781 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tw6vm" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.964838 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-49wzf" podUID="f78e585b-e57b-4681-863c-6489cc654222" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.964859 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-l8jn9" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.964948 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jzzcf" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" Oct 03 13:33:49 crc kubenswrapper[4959]: E1003 13:33:49.964969 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lhm6k" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" Oct 03 13:33:50 crc kubenswrapper[4959]: I1003 13:33:50.967896 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gf4tr" event={"ID":"e3a679d1-7412-4843-998e-8d1c109bf7ec","Type":"ContainerStarted","Data":"75c4641ab8dcd8b25187e71a0f4bc839e888ab9702b767512ef19c18a7a29408"} Oct 03 13:33:50 crc kubenswrapper[4959]: I1003 13:33:50.985803 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gf4tr" podStartSLOduration=179.985780978 podStartE2EDuration="2m59.985780978s" podCreationTimestamp="2025-10-03 13:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:33:50.983765504 +0000 UTC m=+200.187108921" watchObservedRunningTime="2025-10-03 13:33:50.985780978 +0000 UTC m=+200.189124395" Oct 03 13:34:06 crc kubenswrapper[4959]: I1003 13:34:06.045004 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:34:06 crc kubenswrapper[4959]: I1003 13:34:06.045697 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:34:06 crc kubenswrapper[4959]: I1003 13:34:06.045762 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:34:06 crc kubenswrapper[4959]: I1003 13:34:06.046507 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:34:06 crc kubenswrapper[4959]: I1003 13:34:06.046643 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c" gracePeriod=600 Oct 03 13:34:07 crc kubenswrapper[4959]: I1003 13:34:07.058222 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c" exitCode=0 Oct 03 13:34:07 crc kubenswrapper[4959]: I1003 13:34:07.058275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.065790 4959 generic.go:334] "Generic (PLEG): container finished" podID="f78e585b-e57b-4681-863c-6489cc654222" containerID="afa9ffed6ccda4d3fdb7472634a959dc793f25e9aac405efe6cf5e7a7075da05" exitCode=0 Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.065882 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerDied","Data":"afa9ffed6ccda4d3fdb7472634a959dc793f25e9aac405efe6cf5e7a7075da05"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.068249 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerID="d3f07c4ce04ba261147b720d23b65816448cafdead089a6633ed76f0e2ecec02" exitCode=0 Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.068283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerDied","Data":"d3f07c4ce04ba261147b720d23b65816448cafdead089a6633ed76f0e2ecec02"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.071756 4959 generic.go:334] "Generic (PLEG): container finished" podID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerID="f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e" exitCode=0 Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.071809 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerDied","Data":"f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.073868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerStarted","Data":"7610f6452250ff7cb59e3d9751acc6693657b1bd28296e8bdd080b1d37da3fc4"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.077713 4959 generic.go:334] "Generic (PLEG): container finished" podID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerID="dba86c04c000809947af740de074fbbe7f751f4b44a160fc8edf76f194b5010c" exitCode=0 Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.077770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerDied","Data":"dba86c04c000809947af740de074fbbe7f751f4b44a160fc8edf76f194b5010c"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.080752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540"} Oct 03 13:34:08 crc kubenswrapper[4959]: I1003 13:34:08.090316 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerStarted","Data":"8be5dd1f65e8991be6736f109930796a6e91064a5c3f55dc5be4e86ab3689ad9"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.098439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerStarted","Data":"a902d91dae45d540b10fca30d241aa1f9e1f73c4dcb4976aa077129821acfdca"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.102353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerStarted","Data":"6308ee8cf586bb3dde185c4f5bb0a47f065aab512419f0efec0f63cdbe67a7a8"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.106800 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerStarted","Data":"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.109111 4959 generic.go:334] "Generic (PLEG): container finished" podID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerID="7610f6452250ff7cb59e3d9751acc6693657b1bd28296e8bdd080b1d37da3fc4" exitCode=0 Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.109227 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerDied","Data":"7610f6452250ff7cb59e3d9751acc6693657b1bd28296e8bdd080b1d37da3fc4"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.114308 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerStarted","Data":"230f4b3ba2afcb3863bf7068ebc59ff4e70ff78cf323beb146f249d00172eb27"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.118628 4959 generic.go:334] "Generic (PLEG): container finished" podID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerID="72684ee3aacee13f8315a5366d577c842f770128f9ca7a83f621afd36d03add1" exitCode=0 Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.118706 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerDied","Data":"72684ee3aacee13f8315a5366d577c842f770128f9ca7a83f621afd36d03add1"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.124173 4959 generic.go:334] "Generic (PLEG): container finished" podID="0324bc20-8d56-4195-b566-aad202e4fe70" containerID="dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4" exitCode=0 Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.124367 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerDied","Data":"dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.127135 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49wzf" podStartSLOduration=3.9939923200000003 podStartE2EDuration="1m11.127114746s" podCreationTimestamp="2025-10-03 13:32:58 +0000 UTC" firstStartedPulling="2025-10-03 13:33:01.374243242 +0000 UTC m=+150.577586659" lastFinishedPulling="2025-10-03 13:34:08.507365668 +0000 UTC m=+217.710709085" observedRunningTime="2025-10-03 13:34:09.123479389 +0000 UTC m=+218.326822816" watchObservedRunningTime="2025-10-03 13:34:09.127114746 +0000 UTC m=+218.330458173" Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.131364 4959 generic.go:334] "Generic (PLEG): container finished" podID="60a8d300-9415-476c-871b-ef8400c8656c" containerID="8be5dd1f65e8991be6736f109930796a6e91064a5c3f55dc5be4e86ab3689ad9" exitCode=0 Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.135729 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerDied","Data":"8be5dd1f65e8991be6736f109930796a6e91064a5c3f55dc5be4e86ab3689ad9"} Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.177473 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jnggp" podStartSLOduration=3.763012477 podStartE2EDuration="1m11.177394469s" podCreationTimestamp="2025-10-03 13:32:58 +0000 UTC" firstStartedPulling="2025-10-03 13:33:01.378320759 +0000 UTC m=+150.581664176" lastFinishedPulling="2025-10-03 13:34:08.792702751 +0000 UTC m=+217.996046168" observedRunningTime="2025-10-03 13:34:09.156512566 +0000 UTC m=+218.359855983" watchObservedRunningTime="2025-10-03 13:34:09.177394469 +0000 UTC m=+218.380737886" Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.177730 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jzzcf" podStartSLOduration=3.152159853 podStartE2EDuration="1m9.177726049s" podCreationTimestamp="2025-10-03 13:33:00 +0000 UTC" firstStartedPulling="2025-10-03 13:33:02.54654788 +0000 UTC m=+151.749891297" lastFinishedPulling="2025-10-03 13:34:08.572114066 +0000 UTC m=+217.775457493" observedRunningTime="2025-10-03 13:34:09.17551045 +0000 UTC m=+218.378853867" watchObservedRunningTime="2025-10-03 13:34:09.177726049 +0000 UTC m=+218.381069466" Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.243940 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7zsrk" podStartSLOduration=3.097955964 podStartE2EDuration="1m9.243914605s" podCreationTimestamp="2025-10-03 13:33:00 +0000 UTC" firstStartedPulling="2025-10-03 13:33:02.457719214 +0000 UTC m=+151.661062631" lastFinishedPulling="2025-10-03 13:34:08.603677855 +0000 UTC m=+217.807021272" observedRunningTime="2025-10-03 13:34:09.242082757 +0000 UTC m=+218.445426184" watchObservedRunningTime="2025-10-03 13:34:09.243914605 +0000 UTC m=+218.447258022" Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.279088 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:09 crc kubenswrapper[4959]: I1003 13:34:09.279134 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.139609 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerStarted","Data":"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad"} Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.142534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerStarted","Data":"c198d0af284a5c025b501b535de609c1f8525ff372a1c855a8e23c724cd8a7cc"} Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.148657 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerStarted","Data":"ee0a7d0e1803048d688d14d5306455f8217e9c8942b6049d18109a274e762755"} Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.151897 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerStarted","Data":"991d8e800ac19aea50c0e8102e0110dedc1183fc3aa80990cd75589366cd6460"} Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.160339 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l8jn9" podStartSLOduration=2.888184926 podStartE2EDuration="1m9.160319214s" podCreationTimestamp="2025-10-03 13:33:01 +0000 UTC" firstStartedPulling="2025-10-03 13:33:03.502954413 +0000 UTC m=+152.706297830" lastFinishedPulling="2025-10-03 13:34:09.775088701 +0000 UTC m=+218.978432118" observedRunningTime="2025-10-03 13:34:10.156311269 +0000 UTC m=+219.359654706" watchObservedRunningTime="2025-10-03 13:34:10.160319214 +0000 UTC m=+219.363662631" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.173031 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tw6vm" podStartSLOduration=3.168553643 podStartE2EDuration="1m8.173010492s" podCreationTimestamp="2025-10-03 13:33:02 +0000 UTC" firstStartedPulling="2025-10-03 13:33:04.55612346 +0000 UTC m=+153.759466887" lastFinishedPulling="2025-10-03 13:34:09.560580319 +0000 UTC m=+218.763923736" observedRunningTime="2025-10-03 13:34:10.171424069 +0000 UTC m=+219.374767486" watchObservedRunningTime="2025-10-03 13:34:10.173010492 +0000 UTC m=+219.376353909" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.196992 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lhm6k" podStartSLOduration=3.83761957 podStartE2EDuration="1m12.196969028s" podCreationTimestamp="2025-10-03 13:32:58 +0000 UTC" firstStartedPulling="2025-10-03 13:33:01.332167326 +0000 UTC m=+150.535510743" lastFinishedPulling="2025-10-03 13:34:09.691516784 +0000 UTC m=+218.894860201" observedRunningTime="2025-10-03 13:34:10.193252809 +0000 UTC m=+219.396596226" watchObservedRunningTime="2025-10-03 13:34:10.196969028 +0000 UTC m=+219.400312445" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.211629 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kmtqr" podStartSLOduration=3.042309394 podStartE2EDuration="1m11.211607106s" podCreationTimestamp="2025-10-03 13:32:59 +0000 UTC" firstStartedPulling="2025-10-03 13:33:01.39891659 +0000 UTC m=+150.602260007" lastFinishedPulling="2025-10-03 13:34:09.568214302 +0000 UTC m=+218.771557719" observedRunningTime="2025-10-03 13:34:10.209643494 +0000 UTC m=+219.412986911" watchObservedRunningTime="2025-10-03 13:34:10.211607106 +0000 UTC m=+219.414950533" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.426808 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jnggp" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="registry-server" probeResult="failure" output=< Oct 03 13:34:10 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 13:34:10 crc kubenswrapper[4959]: > Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.860558 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.860631 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:10 crc kubenswrapper[4959]: I1003 13:34:10.925855 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:11 crc kubenswrapper[4959]: I1003 13:34:11.308490 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:11 crc kubenswrapper[4959]: I1003 13:34:11.308563 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:11 crc kubenswrapper[4959]: I1003 13:34:11.364954 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:12 crc kubenswrapper[4959]: I1003 13:34:12.221295 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:12 crc kubenswrapper[4959]: I1003 13:34:12.221661 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:12 crc kubenswrapper[4959]: I1003 13:34:12.702838 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:12 crc kubenswrapper[4959]: I1003 13:34:12.703114 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:13 crc kubenswrapper[4959]: I1003 13:34:13.263869 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l8jn9" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="registry-server" probeResult="failure" output=< Oct 03 13:34:13 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 13:34:13 crc kubenswrapper[4959]: > Oct 03 13:34:13 crc kubenswrapper[4959]: I1003 13:34:13.363686 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:34:13 crc kubenswrapper[4959]: I1003 13:34:13.740867 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tw6vm" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="registry-server" probeResult="failure" output=< Oct 03 13:34:13 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 13:34:13 crc kubenswrapper[4959]: > Oct 03 13:34:18 crc kubenswrapper[4959]: I1003 13:34:18.868497 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:18 crc kubenswrapper[4959]: I1003 13:34:18.869157 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:18 crc kubenswrapper[4959]: I1003 13:34:18.909170 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.057471 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.057531 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.091129 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.234779 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.240322 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.369970 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.410032 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.510116 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.510247 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:19 crc kubenswrapper[4959]: I1003 13:34:19.546464 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:20 crc kubenswrapper[4959]: I1003 13:34:20.238226 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:20 crc kubenswrapper[4959]: I1003 13:34:20.897496 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:20 crc kubenswrapper[4959]: I1003 13:34:20.938256 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:34:21 crc kubenswrapper[4959]: I1003 13:34:21.207762 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jnggp" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="registry-server" containerID="cri-o://6308ee8cf586bb3dde185c4f5bb0a47f065aab512419f0efec0f63cdbe67a7a8" gracePeriod=2 Oct 03 13:34:21 crc kubenswrapper[4959]: I1003 13:34:21.346009 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:21 crc kubenswrapper[4959]: I1003 13:34:21.536909 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:34:22 crc kubenswrapper[4959]: I1003 13:34:22.257380 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:22 crc kubenswrapper[4959]: I1003 13:34:22.298727 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:22 crc kubenswrapper[4959]: I1003 13:34:22.749291 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:22 crc kubenswrapper[4959]: I1003 13:34:22.790661 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.219600 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerID="6308ee8cf586bb3dde185c4f5bb0a47f065aab512419f0efec0f63cdbe67a7a8" exitCode=0 Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.219999 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerDied","Data":"6308ee8cf586bb3dde185c4f5bb0a47f065aab512419f0efec0f63cdbe67a7a8"} Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.220405 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kmtqr" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="registry-server" containerID="cri-o://c198d0af284a5c025b501b535de609c1f8525ff372a1c855a8e23c724cd8a7cc" gracePeriod=2 Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.341227 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.341762 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jzzcf" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="registry-server" containerID="cri-o://230f4b3ba2afcb3863bf7068ebc59ff4e70ff78cf323beb146f249d00172eb27" gracePeriod=2 Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.436721 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.578248 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knxkc\" (UniqueName: \"kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc\") pod \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.578356 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities\") pod \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.578427 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content\") pod \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\" (UID: \"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b\") " Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.585302 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities" (OuterVolumeSpecName: "utilities") pod "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" (UID: "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.593843 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc" (OuterVolumeSpecName: "kube-api-access-knxkc") pod "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" (UID: "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b"). InnerVolumeSpecName "kube-api-access-knxkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.680997 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knxkc\" (UniqueName: \"kubernetes.io/projected/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-kube-api-access-knxkc\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.681037 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.682784 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" (UID: "e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:23 crc kubenswrapper[4959]: I1003 13:34:23.781834 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.226483 4959 generic.go:334] "Generic (PLEG): container finished" podID="60a8d300-9415-476c-871b-ef8400c8656c" containerID="c198d0af284a5c025b501b535de609c1f8525ff372a1c855a8e23c724cd8a7cc" exitCode=0 Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.226538 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerDied","Data":"c198d0af284a5c025b501b535de609c1f8525ff372a1c855a8e23c724cd8a7cc"} Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.228911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jnggp" event={"ID":"e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b","Type":"ContainerDied","Data":"d70439f58b96bd366813c5dee4946f692ef8494016978edc1781a54c76ca8470"} Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.228929 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jnggp" Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.229136 4959 scope.go:117] "RemoveContainer" containerID="6308ee8cf586bb3dde185c4f5bb0a47f065aab512419f0efec0f63cdbe67a7a8" Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.230761 4959 generic.go:334] "Generic (PLEG): container finished" podID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerID="230f4b3ba2afcb3863bf7068ebc59ff4e70ff78cf323beb146f249d00172eb27" exitCode=0 Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.230807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerDied","Data":"230f4b3ba2afcb3863bf7068ebc59ff4e70ff78cf323beb146f249d00172eb27"} Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.249535 4959 scope.go:117] "RemoveContainer" containerID="d3f07c4ce04ba261147b720d23b65816448cafdead089a6633ed76f0e2ecec02" Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.253801 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.256774 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jnggp"] Oct 03 13:34:24 crc kubenswrapper[4959]: I1003 13:34:24.265475 4959 scope.go:117] "RemoveContainer" containerID="1452e11e23cf35873af0fc9da5542fa42b4b26f93d3db85ec314a77664781c73" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.413388 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.454484 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.604646 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content\") pod \"60a8d300-9415-476c-871b-ef8400c8656c\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.604779 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content\") pod \"8132bb0b-e372-4e00-93dd-8be8982cad6f\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.604806 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities\") pod \"8132bb0b-e372-4e00-93dd-8be8982cad6f\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.604871 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ngvn\" (UniqueName: \"kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn\") pod \"60a8d300-9415-476c-871b-ef8400c8656c\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.604894 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gn92\" (UniqueName: \"kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92\") pod \"8132bb0b-e372-4e00-93dd-8be8982cad6f\" (UID: \"8132bb0b-e372-4e00-93dd-8be8982cad6f\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.605598 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities" (OuterVolumeSpecName: "utilities") pod "8132bb0b-e372-4e00-93dd-8be8982cad6f" (UID: "8132bb0b-e372-4e00-93dd-8be8982cad6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.605678 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities\") pod \"60a8d300-9415-476c-871b-ef8400c8656c\" (UID: \"60a8d300-9415-476c-871b-ef8400c8656c\") " Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.606326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities" (OuterVolumeSpecName: "utilities") pod "60a8d300-9415-476c-871b-ef8400c8656c" (UID: "60a8d300-9415-476c-871b-ef8400c8656c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.606590 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.606606 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.609559 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92" (OuterVolumeSpecName: "kube-api-access-2gn92") pod "8132bb0b-e372-4e00-93dd-8be8982cad6f" (UID: "8132bb0b-e372-4e00-93dd-8be8982cad6f"). InnerVolumeSpecName "kube-api-access-2gn92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.615844 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn" (OuterVolumeSpecName: "kube-api-access-2ngvn") pod "60a8d300-9415-476c-871b-ef8400c8656c" (UID: "60a8d300-9415-476c-871b-ef8400c8656c"). InnerVolumeSpecName "kube-api-access-2ngvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.621239 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8132bb0b-e372-4e00-93dd-8be8982cad6f" (UID: "8132bb0b-e372-4e00-93dd-8be8982cad6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.652244 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60a8d300-9415-476c-871b-ef8400c8656c" (UID: "60a8d300-9415-476c-871b-ef8400c8656c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.694010 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" path="/var/lib/kubelet/pods/e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b/volumes" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.710810 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ngvn\" (UniqueName: \"kubernetes.io/projected/60a8d300-9415-476c-871b-ef8400c8656c-kube-api-access-2ngvn\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.710878 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gn92\" (UniqueName: \"kubernetes.io/projected/8132bb0b-e372-4e00-93dd-8be8982cad6f-kube-api-access-2gn92\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.710895 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60a8d300-9415-476c-871b-ef8400c8656c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.710909 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8132bb0b-e372-4e00-93dd-8be8982cad6f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.938778 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:34:25 crc kubenswrapper[4959]: I1003 13:34:25.939377 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tw6vm" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="registry-server" containerID="cri-o://991d8e800ac19aea50c0e8102e0110dedc1183fc3aa80990cd75589366cd6460" gracePeriod=2 Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.254171 4959 generic.go:334] "Generic (PLEG): container finished" podID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerID="991d8e800ac19aea50c0e8102e0110dedc1183fc3aa80990cd75589366cd6460" exitCode=0 Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.254224 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerDied","Data":"991d8e800ac19aea50c0e8102e0110dedc1183fc3aa80990cd75589366cd6460"} Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.256352 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmtqr" event={"ID":"60a8d300-9415-476c-871b-ef8400c8656c","Type":"ContainerDied","Data":"dacc5d77bac287313afa9b5fdaface198b330e5e3b284769b768bb536fdadfea"} Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.256391 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmtqr" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.256409 4959 scope.go:117] "RemoveContainer" containerID="c198d0af284a5c025b501b535de609c1f8525ff372a1c855a8e23c724cd8a7cc" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.261577 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzzcf" event={"ID":"8132bb0b-e372-4e00-93dd-8be8982cad6f","Type":"ContainerDied","Data":"b19242379a1d10012c05a01e1d3bc06a8184253f9a913928fcacfb6ad39a45aa"} Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.261690 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzzcf" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.276614 4959 scope.go:117] "RemoveContainer" containerID="8be5dd1f65e8991be6736f109930796a6e91064a5c3f55dc5be4e86ab3689ad9" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.283575 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.293705 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kmtqr"] Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.298707 4959 scope.go:117] "RemoveContainer" containerID="e7c69e414edc1506dd8c7681e6401e6c3e8115dc01645ab8a2829f5970e889de" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.302955 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.305434 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzzcf"] Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.316647 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.316925 4959 scope.go:117] "RemoveContainer" containerID="230f4b3ba2afcb3863bf7068ebc59ff4e70ff78cf323beb146f249d00172eb27" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.335813 4959 scope.go:117] "RemoveContainer" containerID="dba86c04c000809947af740de074fbbe7f751f4b44a160fc8edf76f194b5010c" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.352570 4959 scope.go:117] "RemoveContainer" containerID="6b92671e94ccae1be82d65e5968388ba378139544bfd4e989f3ae22558ae4deb" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.420238 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities\") pod \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.420323 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content\") pod \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.420377 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckpcs\" (UniqueName: \"kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs\") pod \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\" (UID: \"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc\") " Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.421616 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities" (OuterVolumeSpecName: "utilities") pod "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" (UID: "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.425010 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs" (OuterVolumeSpecName: "kube-api-access-ckpcs") pod "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" (UID: "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc"). InnerVolumeSpecName "kube-api-access-ckpcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.500735 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" (UID: "c7a296e7-4d6f-40de-a3f3-567cc0bdedcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.522629 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.522739 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckpcs\" (UniqueName: \"kubernetes.io/projected/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-kube-api-access-ckpcs\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:26 crc kubenswrapper[4959]: I1003 13:34:26.522802 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.270211 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tw6vm" event={"ID":"c7a296e7-4d6f-40de-a3f3-567cc0bdedcc","Type":"ContainerDied","Data":"cdc9bcec9cea9698f8816e0377a87bfacab47369b1eadc8ba3fa830fa5d177be"} Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.270442 4959 scope.go:117] "RemoveContainer" containerID="991d8e800ac19aea50c0e8102e0110dedc1183fc3aa80990cd75589366cd6460" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.270291 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tw6vm" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.288493 4959 scope.go:117] "RemoveContainer" containerID="72684ee3aacee13f8315a5366d577c842f770128f9ca7a83f621afd36d03add1" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.307406 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.309969 4959 scope.go:117] "RemoveContainer" containerID="d0037b2450eb5432cf9e0ea2e7c6099282aab12b9e271c818dc4724352ce1767" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.310475 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tw6vm"] Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.692379 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a8d300-9415-476c-871b-ef8400c8656c" path="/var/lib/kubelet/pods/60a8d300-9415-476c-871b-ef8400c8656c/volumes" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.695148 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" path="/var/lib/kubelet/pods/8132bb0b-e372-4e00-93dd-8be8982cad6f/volumes" Oct 03 13:34:27 crc kubenswrapper[4959]: I1003 13:34:27.695733 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" path="/var/lib/kubelet/pods/c7a296e7-4d6f-40de-a3f3-567cc0bdedcc/volumes" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.392311 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" podUID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" containerName="oauth-openshift" containerID="cri-o://524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1" gracePeriod=15 Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.742533 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773035 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2"] Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773288 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773304 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773318 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f92826-313a-4192-b1f3-479f751ebf20" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773326 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f92826-313a-4192-b1f3-479f751ebf20" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773338 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773346 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773355 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c97e09-88e1-4f2b-a476-19796c3d83f2" containerName="collect-profiles" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773362 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c97e09-88e1-4f2b-a476-19796c3d83f2" containerName="collect-profiles" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773374 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773381 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773393 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773400 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773413 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773421 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773432 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773438 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773446 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773451 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773462 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773468 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773475 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" containerName="oauth-openshift" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773481 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" containerName="oauth-openshift" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773487 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773493 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="extract-utilities" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773502 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773508 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773514 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773520 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773530 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1460159d-ec5f-4b4a-821a-9c0a492b8d0d" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773535 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1460159d-ec5f-4b4a-821a-9c0a492b8d0d" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: E1003 13:34:38.773541 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773547 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="extract-content" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773633 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8132bb0b-e372-4e00-93dd-8be8982cad6f" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773645 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f92826-313a-4192-b1f3-479f751ebf20" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773652 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a296e7-4d6f-40de-a3f3-567cc0bdedcc" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773661 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a8d300-9415-476c-871b-ef8400c8656c" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773670 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1460159d-ec5f-4b4a-821a-9c0a492b8d0d" containerName="pruner" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773682 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6defaa4-2d2b-4c1f-86e9-9b8a6a0ef49b" containerName="registry-server" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773693 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" containerName="oauth-openshift" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.773702 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6c97e09-88e1-4f2b-a476-19796c3d83f2" containerName="collect-profiles" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.774106 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.782558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2"] Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.864355 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pwqb\" (UniqueName: \"kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865390 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865429 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865453 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865469 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865490 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865510 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865527 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865542 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865572 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865605 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865645 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865681 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs\") pod \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\" (UID: \"aa273389-8a40-4e29-b0cd-1be72ac79cb6\") " Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865829 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865869 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865907 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865942 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.865973 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-policies\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866427 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx9rj\" (UniqueName: \"kubernetes.io/projected/41e2a89a-2715-44f8-b39a-5b9151f5c82f-kube-api-access-nx9rj\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866460 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866509 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-dir\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866499 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866590 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866645 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866660 4959 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.866884 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.868908 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.870824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.871147 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb" (OuterVolumeSpecName: "kube-api-access-8pwqb") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "kube-api-access-8pwqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.872149 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.873235 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.873531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.873572 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.875575 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.876518 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.877560 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "aa273389-8a40-4e29-b0cd-1be72ac79cb6" (UID: "aa273389-8a40-4e29-b0cd-1be72ac79cb6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967798 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-dir\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967902 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967930 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.967996 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-dir\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968035 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968108 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968168 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968244 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968276 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968327 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-policies\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx9rj\" (UniqueName: \"kubernetes.io/projected/41e2a89a-2715-44f8-b39a-5b9151f5c82f-kube-api-access-nx9rj\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968414 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968486 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968503 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968517 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968530 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968541 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968554 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pwqb\" (UniqueName: \"kubernetes.io/projected/aa273389-8a40-4e29-b0cd-1be72ac79cb6-kube-api-access-8pwqb\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968566 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968580 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968595 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968607 4959 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aa273389-8a40-4e29-b0cd-1be72ac79cb6-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968620 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.968632 4959 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aa273389-8a40-4e29-b0cd-1be72ac79cb6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.969152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.969493 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.969762 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.969968 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/41e2a89a-2715-44f8-b39a-5b9151f5c82f-audit-policies\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.971296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.971302 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-error\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.972868 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-session\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.972916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.973409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.974153 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-login\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.974173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.974291 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/41e2a89a-2715-44f8-b39a-5b9151f5c82f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:38 crc kubenswrapper[4959]: I1003 13:34:38.984026 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx9rj\" (UniqueName: \"kubernetes.io/projected/41e2a89a-2715-44f8-b39a-5b9151f5c82f-kube-api-access-nx9rj\") pod \"oauth-openshift-54f7c55fd8-xf5x2\" (UID: \"41e2a89a-2715-44f8-b39a-5b9151f5c82f\") " pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.092806 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.277165 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2"] Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.335532 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" containerID="524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1" exitCode=0 Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.335585 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.335578 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" event={"ID":"aa273389-8a40-4e29-b0cd-1be72ac79cb6","Type":"ContainerDied","Data":"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1"} Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.336013 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jqpvt" event={"ID":"aa273389-8a40-4e29-b0cd-1be72ac79cb6","Type":"ContainerDied","Data":"3150e663e7375ed21a6be27f04aa1a4b050b80efcb193a009f90ed3dd0690acb"} Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.336032 4959 scope.go:117] "RemoveContainer" containerID="524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.337457 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" event={"ID":"41e2a89a-2715-44f8-b39a-5b9151f5c82f","Type":"ContainerStarted","Data":"ee41d73eb4f43691c23c613dd0b112ca70c361b9935dcbf852ec37c7696725de"} Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.352185 4959 scope.go:117] "RemoveContainer" containerID="524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1" Oct 03 13:34:39 crc kubenswrapper[4959]: E1003 13:34:39.352672 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1\": container with ID starting with 524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1 not found: ID does not exist" containerID="524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.352721 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1"} err="failed to get container status \"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1\": rpc error: code = NotFound desc = could not find container \"524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1\": container with ID starting with 524d640a9eb2e67172fdd57d14fccb775bf0da33426cf857173134b276714df1 not found: ID does not exist" Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.364765 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.367853 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jqpvt"] Oct 03 13:34:39 crc kubenswrapper[4959]: I1003 13:34:39.693758 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa273389-8a40-4e29-b0cd-1be72ac79cb6" path="/var/lib/kubelet/pods/aa273389-8a40-4e29-b0cd-1be72ac79cb6/volumes" Oct 03 13:34:40 crc kubenswrapper[4959]: I1003 13:34:40.347766 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" event={"ID":"41e2a89a-2715-44f8-b39a-5b9151f5c82f","Type":"ContainerStarted","Data":"7a1752a149025f78b2dfc5b3fb95f4df986a0ffef43760309fc28e24238c5a3b"} Oct 03 13:34:40 crc kubenswrapper[4959]: I1003 13:34:40.348083 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:40 crc kubenswrapper[4959]: I1003 13:34:40.353692 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" Oct 03 13:34:40 crc kubenswrapper[4959]: I1003 13:34:40.373770 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-54f7c55fd8-xf5x2" podStartSLOduration=27.373750493 podStartE2EDuration="27.373750493s" podCreationTimestamp="2025-10-03 13:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:34:40.370273661 +0000 UTC m=+249.573617098" watchObservedRunningTime="2025-10-03 13:34:40.373750493 +0000 UTC m=+249.577093920" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.262904 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.264613 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-49wzf" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="registry-server" containerID="cri-o://a902d91dae45d540b10fca30d241aa1f9e1f73c4dcb4976aa077129821acfdca" gracePeriod=30 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.273729 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.273998 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lhm6k" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="registry-server" containerID="cri-o://ee0a7d0e1803048d688d14d5306455f8217e9c8942b6049d18109a274e762755" gracePeriod=30 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.280436 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.280634 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" containerID="cri-o://a43cf8ce9f273837c4077f34b5d050ff93fc048d9b8226694dd43c05d99fc575" gracePeriod=30 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.284936 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.285246 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7zsrk" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="registry-server" containerID="cri-o://f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313" gracePeriod=30 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.290422 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.290705 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l8jn9" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="registry-server" containerID="cri-o://c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad" gracePeriod=30 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.302991 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qxbr"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.303616 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.311034 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qxbr"] Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.345684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.345729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v2zp\" (UniqueName: \"kubernetes.io/projected/0ad50895-bab1-463a-97ac-3765cbf374bc-kube-api-access-2v2zp\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.345853 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.409665 4959 generic.go:334] "Generic (PLEG): container finished" podID="f78e585b-e57b-4681-863c-6489cc654222" containerID="a902d91dae45d540b10fca30d241aa1f9e1f73c4dcb4976aa077129821acfdca" exitCode=0 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.410009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerDied","Data":"a902d91dae45d540b10fca30d241aa1f9e1f73c4dcb4976aa077129821acfdca"} Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.413763 4959 generic.go:334] "Generic (PLEG): container finished" podID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerID="ee0a7d0e1803048d688d14d5306455f8217e9c8942b6049d18109a274e762755" exitCode=0 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.413805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerDied","Data":"ee0a7d0e1803048d688d14d5306455f8217e9c8942b6049d18109a274e762755"} Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.417478 4959 generic.go:334] "Generic (PLEG): container finished" podID="d8713790-50b4-4939-b47d-27927aefcddc" containerID="a43cf8ce9f273837c4077f34b5d050ff93fc048d9b8226694dd43c05d99fc575" exitCode=0 Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.417510 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" event={"ID":"d8713790-50b4-4939-b47d-27927aefcddc","Type":"ContainerDied","Data":"a43cf8ce9f273837c4077f34b5d050ff93fc048d9b8226694dd43c05d99fc575"} Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.446883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.446939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v2zp\" (UniqueName: \"kubernetes.io/projected/0ad50895-bab1-463a-97ac-3765cbf374bc-kube-api-access-2v2zp\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.446983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.448617 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.453776 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0ad50895-bab1-463a-97ac-3765cbf374bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.468008 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v2zp\" (UniqueName: \"kubernetes.io/projected/0ad50895-bab1-463a-97ac-3765cbf374bc-kube-api-access-2v2zp\") pod \"marketplace-operator-79b997595-9qxbr\" (UID: \"0ad50895-bab1-463a-97ac-3765cbf374bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.645088 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.651070 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.727338 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.730990 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751075 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca\") pod \"d8713790-50b4-4939-b47d-27927aefcddc\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751142 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content\") pod \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751168 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content\") pod \"f78e585b-e57b-4681-863c-6489cc654222\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751230 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjrsz\" (UniqueName: \"kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz\") pod \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751266 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics\") pod \"d8713790-50b4-4939-b47d-27927aefcddc\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751295 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities\") pod \"f78e585b-e57b-4681-863c-6489cc654222\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751319 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6jrn\" (UniqueName: \"kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn\") pod \"f78e585b-e57b-4681-863c-6489cc654222\" (UID: \"f78e585b-e57b-4681-863c-6489cc654222\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751345 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrdhw\" (UniqueName: \"kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw\") pod \"d8713790-50b4-4939-b47d-27927aefcddc\" (UID: \"d8713790-50b4-4939-b47d-27927aefcddc\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.751399 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities\") pod \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\" (UID: \"588d8af1-ff2a-44de-ad5f-2fb64b934cef\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.752687 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d8713790-50b4-4939-b47d-27927aefcddc" (UID: "d8713790-50b4-4939-b47d-27927aefcddc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.759536 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities" (OuterVolumeSpecName: "utilities") pod "588d8af1-ff2a-44de-ad5f-2fb64b934cef" (UID: "588d8af1-ff2a-44de-ad5f-2fb64b934cef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.760653 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz" (OuterVolumeSpecName: "kube-api-access-mjrsz") pod "588d8af1-ff2a-44de-ad5f-2fb64b934cef" (UID: "588d8af1-ff2a-44de-ad5f-2fb64b934cef"). InnerVolumeSpecName "kube-api-access-mjrsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.761027 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities" (OuterVolumeSpecName: "utilities") pod "f78e585b-e57b-4681-863c-6489cc654222" (UID: "f78e585b-e57b-4681-863c-6489cc654222"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.760620 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn" (OuterVolumeSpecName: "kube-api-access-b6jrn") pod "f78e585b-e57b-4681-863c-6489cc654222" (UID: "f78e585b-e57b-4681-863c-6489cc654222"). InnerVolumeSpecName "kube-api-access-b6jrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.764272 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.769028 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d8713790-50b4-4939-b47d-27927aefcddc" (UID: "d8713790-50b4-4939-b47d-27927aefcddc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.774845 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.779471 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw" (OuterVolumeSpecName: "kube-api-access-qrdhw") pod "d8713790-50b4-4939-b47d-27927aefcddc" (UID: "d8713790-50b4-4939-b47d-27927aefcddc"). InnerVolumeSpecName "kube-api-access-qrdhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.845040 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f78e585b-e57b-4681-863c-6489cc654222" (UID: "f78e585b-e57b-4681-863c-6489cc654222"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.845667 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "588d8af1-ff2a-44de-ad5f-2fb64b934cef" (UID: "588d8af1-ff2a-44de-ad5f-2fb64b934cef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.851914 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities\") pod \"0324bc20-8d56-4195-b566-aad202e4fe70\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.851957 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content\") pod \"0324bc20-8d56-4195-b566-aad202e4fe70\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.851986 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8gpv\" (UniqueName: \"kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv\") pod \"0324bc20-8d56-4195-b566-aad202e4fe70\" (UID: \"0324bc20-8d56-4195-b566-aad202e4fe70\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content\") pod \"1bee80ae-896e-4d7d-885d-e1bf8280db86\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852033 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities\") pod \"1bee80ae-896e-4d7d-885d-e1bf8280db86\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852072 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwgpk\" (UniqueName: \"kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk\") pod \"1bee80ae-896e-4d7d-885d-e1bf8280db86\" (UID: \"1bee80ae-896e-4d7d-885d-e1bf8280db86\") " Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852268 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852287 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852300 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/588d8af1-ff2a-44de-ad5f-2fb64b934cef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852311 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852320 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjrsz\" (UniqueName: \"kubernetes.io/projected/588d8af1-ff2a-44de-ad5f-2fb64b934cef-kube-api-access-mjrsz\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852332 4959 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8713790-50b4-4939-b47d-27927aefcddc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852344 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78e585b-e57b-4681-863c-6489cc654222-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852355 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6jrn\" (UniqueName: \"kubernetes.io/projected/f78e585b-e57b-4681-863c-6489cc654222-kube-api-access-b6jrn\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.852365 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrdhw\" (UniqueName: \"kubernetes.io/projected/d8713790-50b4-4939-b47d-27927aefcddc-kube-api-access-qrdhw\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.853664 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities" (OuterVolumeSpecName: "utilities") pod "0324bc20-8d56-4195-b566-aad202e4fe70" (UID: "0324bc20-8d56-4195-b566-aad202e4fe70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.853784 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities" (OuterVolumeSpecName: "utilities") pod "1bee80ae-896e-4d7d-885d-e1bf8280db86" (UID: "1bee80ae-896e-4d7d-885d-e1bf8280db86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.855532 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk" (OuterVolumeSpecName: "kube-api-access-hwgpk") pod "1bee80ae-896e-4d7d-885d-e1bf8280db86" (UID: "1bee80ae-896e-4d7d-885d-e1bf8280db86"). InnerVolumeSpecName "kube-api-access-hwgpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.855925 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv" (OuterVolumeSpecName: "kube-api-access-w8gpv") pod "0324bc20-8d56-4195-b566-aad202e4fe70" (UID: "0324bc20-8d56-4195-b566-aad202e4fe70"). InnerVolumeSpecName "kube-api-access-w8gpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.874641 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bee80ae-896e-4d7d-885d-e1bf8280db86" (UID: "1bee80ae-896e-4d7d-885d-e1bf8280db86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.948130 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0324bc20-8d56-4195-b566-aad202e4fe70" (UID: "0324bc20-8d56-4195-b566-aad202e4fe70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960846 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960910 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwgpk\" (UniqueName: \"kubernetes.io/projected/1bee80ae-896e-4d7d-885d-e1bf8280db86-kube-api-access-hwgpk\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960926 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960938 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0324bc20-8d56-4195-b566-aad202e4fe70-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960949 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8gpv\" (UniqueName: \"kubernetes.io/projected/0324bc20-8d56-4195-b566-aad202e4fe70-kube-api-access-w8gpv\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:52 crc kubenswrapper[4959]: I1003 13:34:52.960961 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bee80ae-896e-4d7d-885d-e1bf8280db86-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.188524 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qxbr"] Oct 03 13:34:53 crc kubenswrapper[4959]: W1003 13:34:53.195443 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ad50895_bab1_463a_97ac_3765cbf374bc.slice/crio-28f9fd84e286e48a45d33dd2bbc3fb69f2b8597c2877a738dd19f5354f72693a WatchSource:0}: Error finding container 28f9fd84e286e48a45d33dd2bbc3fb69f2b8597c2877a738dd19f5354f72693a: Status 404 returned error can't find the container with id 28f9fd84e286e48a45d33dd2bbc3fb69f2b8597c2877a738dd19f5354f72693a Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.431302 4959 generic.go:334] "Generic (PLEG): container finished" podID="0324bc20-8d56-4195-b566-aad202e4fe70" containerID="c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad" exitCode=0 Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.431545 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerDied","Data":"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.431840 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8jn9" event={"ID":"0324bc20-8d56-4195-b566-aad202e4fe70","Type":"ContainerDied","Data":"56620885646bb078d74a485bfe68c8c6fc87757377e311ba75d96fcc43080846"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.431869 4959 scope.go:117] "RemoveContainer" containerID="c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.431595 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8jn9" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.433176 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" event={"ID":"0ad50895-bab1-463a-97ac-3765cbf374bc","Type":"ContainerStarted","Data":"f6b2369b36a65220ecf37bb72fee5c231aca6c0787f3e7e18a52cbcb2c26a8e0"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.433233 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" event={"ID":"0ad50895-bab1-463a-97ac-3765cbf374bc","Type":"ContainerStarted","Data":"28f9fd84e286e48a45d33dd2bbc3fb69f2b8597c2877a738dd19f5354f72693a"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.433442 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.434942 4959 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9qxbr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.434988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49wzf" event={"ID":"f78e585b-e57b-4681-863c-6489cc654222","Type":"ContainerDied","Data":"2fe743f297112f8470905b13cc1d04885f88857862619de4cf7bc20b95956444"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.435009 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" podUID="0ad50895-bab1-463a-97ac-3765cbf374bc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.435076 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49wzf" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.442536 4959 generic.go:334] "Generic (PLEG): container finished" podID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerID="f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313" exitCode=0 Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.442647 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerDied","Data":"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.442676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zsrk" event={"ID":"1bee80ae-896e-4d7d-885d-e1bf8280db86","Type":"ContainerDied","Data":"40a1d1e61ce721e87d4c5050ef2c01a9abd29f82f7601e1d7a7abdc83a706fb4"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.443007 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zsrk" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.445920 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lhm6k" event={"ID":"588d8af1-ff2a-44de-ad5f-2fb64b934cef","Type":"ContainerDied","Data":"3041b63fcd4dafb961171958912d72d281206f73b90651fbfb4c686d95d284fe"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.446023 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lhm6k" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.454405 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" event={"ID":"d8713790-50b4-4939-b47d-27927aefcddc","Type":"ContainerDied","Data":"d599444c85f1ea0ea59c0b2873e643f0d0be3b503fdddff287add82ea1b0ee2b"} Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.454503 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gspcs" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.469660 4959 scope.go:117] "RemoveContainer" containerID="dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.476976 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" podStartSLOduration=1.476956126 podStartE2EDuration="1.476956126s" podCreationTimestamp="2025-10-03 13:34:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:34:53.452116869 +0000 UTC m=+262.655460306" watchObservedRunningTime="2025-10-03 13:34:53.476956126 +0000 UTC m=+262.680299543" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.478164 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.482413 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l8jn9"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.486946 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.491463 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lhm6k"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.502555 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.505605 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zsrk"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.508454 4959 scope.go:117] "RemoveContainer" containerID="15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.526338 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.526388 4959 scope.go:117] "RemoveContainer" containerID="c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.526862 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad\": container with ID starting with c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad not found: ID does not exist" containerID="c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.526907 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad"} err="failed to get container status \"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad\": rpc error: code = NotFound desc = could not find container \"c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad\": container with ID starting with c65a77e636c26e2ed0654f24db33fe7ceb0b03473d274b51d9633b0f00ed61ad not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.526943 4959 scope.go:117] "RemoveContainer" containerID="dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.527308 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4\": container with ID starting with dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4 not found: ID does not exist" containerID="dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.527334 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4"} err="failed to get container status \"dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4\": rpc error: code = NotFound desc = could not find container \"dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4\": container with ID starting with dc58061695f0f70495aa3a9a954d9223ec4243e4ef95e2f4c48c4477bc795bc4 not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.527352 4959 scope.go:117] "RemoveContainer" containerID="15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.527565 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212\": container with ID starting with 15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212 not found: ID does not exist" containerID="15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.527584 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212"} err="failed to get container status \"15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212\": rpc error: code = NotFound desc = could not find container \"15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212\": container with ID starting with 15a4393f576bf57b067e4435d8f9c4e61beabe5a74665d01f752fec425db9212 not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.527601 4959 scope.go:117] "RemoveContainer" containerID="a902d91dae45d540b10fca30d241aa1f9e1f73c4dcb4976aa077129821acfdca" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.529390 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gspcs"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.544711 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.544770 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-49wzf"] Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.553698 4959 scope.go:117] "RemoveContainer" containerID="afa9ffed6ccda4d3fdb7472634a959dc793f25e9aac405efe6cf5e7a7075da05" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.592590 4959 scope.go:117] "RemoveContainer" containerID="40a4a7dac42e52f655b25a9cfb8f33dda11b5047745922bd962c00664a2d6138" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.622069 4959 scope.go:117] "RemoveContainer" containerID="f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.642515 4959 scope.go:117] "RemoveContainer" containerID="f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.657850 4959 scope.go:117] "RemoveContainer" containerID="1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.678290 4959 scope.go:117] "RemoveContainer" containerID="f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.678788 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313\": container with ID starting with f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313 not found: ID does not exist" containerID="f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.678829 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313"} err="failed to get container status \"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313\": rpc error: code = NotFound desc = could not find container \"f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313\": container with ID starting with f80e1d427a45e8b27e71aa09ae85cce7d82ee798a1a1de261bc9a9b066531313 not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.678856 4959 scope.go:117] "RemoveContainer" containerID="f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.679331 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e\": container with ID starting with f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e not found: ID does not exist" containerID="f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.679363 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e"} err="failed to get container status \"f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e\": rpc error: code = NotFound desc = could not find container \"f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e\": container with ID starting with f3e7447b7772edb62e9a6c089e500c280438440069f4fe0ced0b059471d2f70e not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.679387 4959 scope.go:117] "RemoveContainer" containerID="1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a" Oct 03 13:34:53 crc kubenswrapper[4959]: E1003 13:34:53.680021 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a\": container with ID starting with 1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a not found: ID does not exist" containerID="1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.680051 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a"} err="failed to get container status \"1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a\": rpc error: code = NotFound desc = could not find container \"1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a\": container with ID starting with 1b218603ad80212e7354f6d4eb7bf6c36325c7ecdf2b8ab155684bc0c9485d6a not found: ID does not exist" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.680069 4959 scope.go:117] "RemoveContainer" containerID="ee0a7d0e1803048d688d14d5306455f8217e9c8942b6049d18109a274e762755" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.690879 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" path="/var/lib/kubelet/pods/0324bc20-8d56-4195-b566-aad202e4fe70/volumes" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.691553 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" path="/var/lib/kubelet/pods/1bee80ae-896e-4d7d-885d-e1bf8280db86/volumes" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.692131 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" path="/var/lib/kubelet/pods/588d8af1-ff2a-44de-ad5f-2fb64b934cef/volumes" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.693239 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8713790-50b4-4939-b47d-27927aefcddc" path="/var/lib/kubelet/pods/d8713790-50b4-4939-b47d-27927aefcddc/volumes" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.693668 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78e585b-e57b-4681-863c-6489cc654222" path="/var/lib/kubelet/pods/f78e585b-e57b-4681-863c-6489cc654222/volumes" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.697449 4959 scope.go:117] "RemoveContainer" containerID="7610f6452250ff7cb59e3d9751acc6693657b1bd28296e8bdd080b1d37da3fc4" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.712785 4959 scope.go:117] "RemoveContainer" containerID="d56705f092f97d78c98817fb48e3cacabedde398f7c7f20d821ace3c14127a95" Oct 03 13:34:53 crc kubenswrapper[4959]: I1003 13:34:53.730538 4959 scope.go:117] "RemoveContainer" containerID="a43cf8ce9f273837c4077f34b5d050ff93fc048d9b8226694dd43c05d99fc575" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.468682 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9qxbr" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.607867 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9h98"] Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608085 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608097 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608112 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608118 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608127 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608132 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608141 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608147 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608155 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608160 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608170 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608178 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608183 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608204 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608213 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608220 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608228 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608233 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="extract-utilities" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608240 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608245 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608253 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608258 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608266 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608271 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: E1003 13:34:54.608281 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608286 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="extract-content" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608369 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bee80ae-896e-4d7d-885d-e1bf8280db86" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608381 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78e585b-e57b-4681-863c-6489cc654222" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608388 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0324bc20-8d56-4195-b566-aad202e4fe70" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608393 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8713790-50b4-4939-b47d-27927aefcddc" containerName="marketplace-operator" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.608401 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="588d8af1-ff2a-44de-ad5f-2fb64b934cef" containerName="registry-server" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.609363 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.611828 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.634950 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9h98"] Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.680775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-utilities\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.680844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m95sl\" (UniqueName: \"kubernetes.io/projected/07334ce0-e2f7-44c8-91d4-76fd59a7e540-kube-api-access-m95sl\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.680960 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-catalog-content\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.782397 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-utilities\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.782452 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m95sl\" (UniqueName: \"kubernetes.io/projected/07334ce0-e2f7-44c8-91d4-76fd59a7e540-kube-api-access-m95sl\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.782515 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-catalog-content\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.783057 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-catalog-content\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.783064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07334ce0-e2f7-44c8-91d4-76fd59a7e540-utilities\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.804415 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m95sl\" (UniqueName: \"kubernetes.io/projected/07334ce0-e2f7-44c8-91d4-76fd59a7e540-kube-api-access-m95sl\") pod \"redhat-marketplace-v9h98\" (UID: \"07334ce0-e2f7-44c8-91d4-76fd59a7e540\") " pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:54 crc kubenswrapper[4959]: I1003 13:34:54.937346 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.115148 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9h98"] Oct 03 13:34:55 crc kubenswrapper[4959]: W1003 13:34:55.125102 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07334ce0_e2f7_44c8_91d4_76fd59a7e540.slice/crio-20c2da2c1dcc82cbbc16ab094867579cf0b8d51e131eb7f0f47a916ab0bb8e2d WatchSource:0}: Error finding container 20c2da2c1dcc82cbbc16ab094867579cf0b8d51e131eb7f0f47a916ab0bb8e2d: Status 404 returned error can't find the container with id 20c2da2c1dcc82cbbc16ab094867579cf0b8d51e131eb7f0f47a916ab0bb8e2d Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.470707 4959 generic.go:334] "Generic (PLEG): container finished" podID="07334ce0-e2f7-44c8-91d4-76fd59a7e540" containerID="04fad41fca03bc4180dabacafa828ea3faac96991ae631d6322aba8c3e4c7713" exitCode=0 Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.471210 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9h98" event={"ID":"07334ce0-e2f7-44c8-91d4-76fd59a7e540","Type":"ContainerDied","Data":"04fad41fca03bc4180dabacafa828ea3faac96991ae631d6322aba8c3e4c7713"} Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.471245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9h98" event={"ID":"07334ce0-e2f7-44c8-91d4-76fd59a7e540","Type":"ContainerStarted","Data":"20c2da2c1dcc82cbbc16ab094867579cf0b8d51e131eb7f0f47a916ab0bb8e2d"} Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.611877 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8b59q"] Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.613350 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.615296 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.628079 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8b59q"] Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.698676 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-catalog-content\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.698754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57cr4\" (UniqueName: \"kubernetes.io/projected/29d71df3-b428-4e3f-bf68-ec8c5c60befd-kube-api-access-57cr4\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.698918 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-utilities\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.800034 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-utilities\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.800096 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-catalog-content\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.800115 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57cr4\" (UniqueName: \"kubernetes.io/projected/29d71df3-b428-4e3f-bf68-ec8c5c60befd-kube-api-access-57cr4\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.800676 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-catalog-content\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.800939 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29d71df3-b428-4e3f-bf68-ec8c5c60befd-utilities\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.820814 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57cr4\" (UniqueName: \"kubernetes.io/projected/29d71df3-b428-4e3f-bf68-ec8c5c60befd-kube-api-access-57cr4\") pod \"certified-operators-8b59q\" (UID: \"29d71df3-b428-4e3f-bf68-ec8c5c60befd\") " pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:55 crc kubenswrapper[4959]: I1003 13:34:55.930949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.312038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8b59q"] Oct 03 13:34:56 crc kubenswrapper[4959]: W1003 13:34:56.317338 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29d71df3_b428_4e3f_bf68_ec8c5c60befd.slice/crio-b6bb185abe50191c67c2a3bcde492c5f562640ce68077505076ef960ac5ac1eb WatchSource:0}: Error finding container b6bb185abe50191c67c2a3bcde492c5f562640ce68077505076ef960ac5ac1eb: Status 404 returned error can't find the container with id b6bb185abe50191c67c2a3bcde492c5f562640ce68077505076ef960ac5ac1eb Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.476667 4959 generic.go:334] "Generic (PLEG): container finished" podID="07334ce0-e2f7-44c8-91d4-76fd59a7e540" containerID="defdc0ba687b1bfe16d6c6d05e54e0ad5af3221be7aef852525e87ed7ae43d0d" exitCode=0 Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.476732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9h98" event={"ID":"07334ce0-e2f7-44c8-91d4-76fd59a7e540","Type":"ContainerDied","Data":"defdc0ba687b1bfe16d6c6d05e54e0ad5af3221be7aef852525e87ed7ae43d0d"} Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.480397 4959 generic.go:334] "Generic (PLEG): container finished" podID="29d71df3-b428-4e3f-bf68-ec8c5c60befd" containerID="8ebbfa045ac965f7474c553f88c534f15f39531be748860873bcada2f39080c8" exitCode=0 Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.480439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8b59q" event={"ID":"29d71df3-b428-4e3f-bf68-ec8c5c60befd","Type":"ContainerDied","Data":"8ebbfa045ac965f7474c553f88c534f15f39531be748860873bcada2f39080c8"} Oct 03 13:34:56 crc kubenswrapper[4959]: I1003 13:34:56.480603 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8b59q" event={"ID":"29d71df3-b428-4e3f-bf68-ec8c5c60befd","Type":"ContainerStarted","Data":"b6bb185abe50191c67c2a3bcde492c5f562640ce68077505076ef960ac5ac1eb"} Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.009963 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5plkc"] Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.011657 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.014907 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.018113 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5plkc"] Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.117637 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-utilities\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.117714 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2cf8\" (UniqueName: \"kubernetes.io/projected/015eb8ae-9d78-41ed-b557-7cc23c222968-kube-api-access-t2cf8\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.117733 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-catalog-content\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.218549 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-catalog-content\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.218627 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-utilities\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.218701 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2cf8\" (UniqueName: \"kubernetes.io/projected/015eb8ae-9d78-41ed-b557-7cc23c222968-kube-api-access-t2cf8\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.219019 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-catalog-content\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.219128 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/015eb8ae-9d78-41ed-b557-7cc23c222968-utilities\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.238818 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2cf8\" (UniqueName: \"kubernetes.io/projected/015eb8ae-9d78-41ed-b557-7cc23c222968-kube-api-access-t2cf8\") pod \"redhat-operators-5plkc\" (UID: \"015eb8ae-9d78-41ed-b557-7cc23c222968\") " pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.348770 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.491908 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9h98" event={"ID":"07334ce0-e2f7-44c8-91d4-76fd59a7e540","Type":"ContainerStarted","Data":"66d929399d32477aeb97a551a2bcd08446c81a8e8f5496fe99ad7521da65e2a8"} Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.516718 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9h98" podStartSLOduration=2.047654057 podStartE2EDuration="3.516693974s" podCreationTimestamp="2025-10-03 13:34:54 +0000 UTC" firstStartedPulling="2025-10-03 13:34:55.472969863 +0000 UTC m=+264.676313290" lastFinishedPulling="2025-10-03 13:34:56.94200979 +0000 UTC m=+266.145353207" observedRunningTime="2025-10-03 13:34:57.508025824 +0000 UTC m=+266.711369261" watchObservedRunningTime="2025-10-03 13:34:57.516693974 +0000 UTC m=+266.720037391" Oct 03 13:34:57 crc kubenswrapper[4959]: I1003 13:34:57.582214 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5plkc"] Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.010041 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4d9lm"] Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.011452 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.013925 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.022390 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d9lm"] Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.133628 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frq6\" (UniqueName: \"kubernetes.io/projected/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-kube-api-access-4frq6\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.133911 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-catalog-content\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.133976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-utilities\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.235150 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-utilities\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.235233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frq6\" (UniqueName: \"kubernetes.io/projected/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-kube-api-access-4frq6\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.235252 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-catalog-content\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.235719 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-catalog-content\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.235963 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-utilities\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.256068 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frq6\" (UniqueName: \"kubernetes.io/projected/16f537b4-e7f5-4c35-9ea9-95d20adbfbab-kube-api-access-4frq6\") pod \"community-operators-4d9lm\" (UID: \"16f537b4-e7f5-4c35-9ea9-95d20adbfbab\") " pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.330184 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.496692 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d9lm"] Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.500238 4959 generic.go:334] "Generic (PLEG): container finished" podID="015eb8ae-9d78-41ed-b557-7cc23c222968" containerID="897140080b3f3c34429586f1d26908ba4959a330ea1c735f37be2a5ec3a948ae" exitCode=0 Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.500337 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5plkc" event={"ID":"015eb8ae-9d78-41ed-b557-7cc23c222968","Type":"ContainerDied","Data":"897140080b3f3c34429586f1d26908ba4959a330ea1c735f37be2a5ec3a948ae"} Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.500368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5plkc" event={"ID":"015eb8ae-9d78-41ed-b557-7cc23c222968","Type":"ContainerStarted","Data":"bed98ee7d05fad852c66c674c4f3850f60dc7b1646cbd5583bab5b337bec43cb"} Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.515029 4959 generic.go:334] "Generic (PLEG): container finished" podID="29d71df3-b428-4e3f-bf68-ec8c5c60befd" containerID="63c18dafb97cf4937e003e265b817fd01ac6982a4f0045e6f7a287d3466bd902" exitCode=0 Oct 03 13:34:58 crc kubenswrapper[4959]: I1003 13:34:58.515672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8b59q" event={"ID":"29d71df3-b428-4e3f-bf68-ec8c5c60befd","Type":"ContainerDied","Data":"63c18dafb97cf4937e003e265b817fd01ac6982a4f0045e6f7a287d3466bd902"} Oct 03 13:34:59 crc kubenswrapper[4959]: I1003 13:34:59.520901 4959 generic.go:334] "Generic (PLEG): container finished" podID="16f537b4-e7f5-4c35-9ea9-95d20adbfbab" containerID="faa8fd3028164150135a1822497998445e218e939aff7d1c05c763a2b7cee9f3" exitCode=0 Oct 03 13:34:59 crc kubenswrapper[4959]: I1003 13:34:59.521070 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d9lm" event={"ID":"16f537b4-e7f5-4c35-9ea9-95d20adbfbab","Type":"ContainerDied","Data":"faa8fd3028164150135a1822497998445e218e939aff7d1c05c763a2b7cee9f3"} Oct 03 13:34:59 crc kubenswrapper[4959]: I1003 13:34:59.521473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d9lm" event={"ID":"16f537b4-e7f5-4c35-9ea9-95d20adbfbab","Type":"ContainerStarted","Data":"7946e0464cfcfd5dac5e2fcde76c91eb42b2ecbad3628bb77e7d8fc2fb6b8eb7"} Oct 03 13:34:59 crc kubenswrapper[4959]: I1003 13:34:59.530949 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8b59q" event={"ID":"29d71df3-b428-4e3f-bf68-ec8c5c60befd","Type":"ContainerStarted","Data":"556c0a9d06bef050c466ab2b83575ef39a9b56b94fcb4859befcfd4a7e03a4ee"} Oct 03 13:34:59 crc kubenswrapper[4959]: I1003 13:34:59.558340 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8b59q" podStartSLOduration=1.861036538 podStartE2EDuration="4.558284167s" podCreationTimestamp="2025-10-03 13:34:55 +0000 UTC" firstStartedPulling="2025-10-03 13:34:56.481566018 +0000 UTC m=+265.684909435" lastFinishedPulling="2025-10-03 13:34:59.178813647 +0000 UTC m=+268.382157064" observedRunningTime="2025-10-03 13:34:59.557055774 +0000 UTC m=+268.760399191" watchObservedRunningTime="2025-10-03 13:34:59.558284167 +0000 UTC m=+268.761627584" Oct 03 13:35:01 crc kubenswrapper[4959]: I1003 13:35:01.540723 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5plkc" event={"ID":"015eb8ae-9d78-41ed-b557-7cc23c222968","Type":"ContainerStarted","Data":"24ece7a24390a9e8bcf80b731f369763e86a6e0d6c5f37c9b99fc08decbf7364"} Oct 03 13:35:01 crc kubenswrapper[4959]: I1003 13:35:01.543746 4959 generic.go:334] "Generic (PLEG): container finished" podID="16f537b4-e7f5-4c35-9ea9-95d20adbfbab" containerID="084bc77c8fbac29931e03e11d45c8993694f878c0c9d3252b3f716d45e5fafa2" exitCode=0 Oct 03 13:35:01 crc kubenswrapper[4959]: I1003 13:35:01.543791 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d9lm" event={"ID":"16f537b4-e7f5-4c35-9ea9-95d20adbfbab","Type":"ContainerDied","Data":"084bc77c8fbac29931e03e11d45c8993694f878c0c9d3252b3f716d45e5fafa2"} Oct 03 13:35:02 crc kubenswrapper[4959]: I1003 13:35:02.552315 4959 generic.go:334] "Generic (PLEG): container finished" podID="015eb8ae-9d78-41ed-b557-7cc23c222968" containerID="24ece7a24390a9e8bcf80b731f369763e86a6e0d6c5f37c9b99fc08decbf7364" exitCode=0 Oct 03 13:35:02 crc kubenswrapper[4959]: I1003 13:35:02.552528 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5plkc" event={"ID":"015eb8ae-9d78-41ed-b557-7cc23c222968","Type":"ContainerDied","Data":"24ece7a24390a9e8bcf80b731f369763e86a6e0d6c5f37c9b99fc08decbf7364"} Oct 03 13:35:02 crc kubenswrapper[4959]: I1003 13:35:02.558249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d9lm" event={"ID":"16f537b4-e7f5-4c35-9ea9-95d20adbfbab","Type":"ContainerStarted","Data":"41d0a72479e2b34159a2ec5653825c5f67d53bced72a390e3c33fba3e221f1c3"} Oct 03 13:35:03 crc kubenswrapper[4959]: I1003 13:35:03.565360 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5plkc" event={"ID":"015eb8ae-9d78-41ed-b557-7cc23c222968","Type":"ContainerStarted","Data":"d016892fd5617a7eb21d9dc410f8f52cd44628fdfc7a03d9e92ef70bb91ab19a"} Oct 03 13:35:03 crc kubenswrapper[4959]: I1003 13:35:03.580244 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5plkc" podStartSLOduration=2.955009456 podStartE2EDuration="7.580227224s" podCreationTimestamp="2025-10-03 13:34:56 +0000 UTC" firstStartedPulling="2025-10-03 13:34:58.503036688 +0000 UTC m=+267.706380105" lastFinishedPulling="2025-10-03 13:35:03.128254456 +0000 UTC m=+272.331597873" observedRunningTime="2025-10-03 13:35:03.579615147 +0000 UTC m=+272.782958574" watchObservedRunningTime="2025-10-03 13:35:03.580227224 +0000 UTC m=+272.783570641" Oct 03 13:35:03 crc kubenswrapper[4959]: I1003 13:35:03.581732 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4d9lm" podStartSLOduration=4.065887894 podStartE2EDuration="6.581726873s" podCreationTimestamp="2025-10-03 13:34:57 +0000 UTC" firstStartedPulling="2025-10-03 13:34:59.526695662 +0000 UTC m=+268.730039079" lastFinishedPulling="2025-10-03 13:35:02.042534641 +0000 UTC m=+271.245878058" observedRunningTime="2025-10-03 13:35:02.592944024 +0000 UTC m=+271.796287441" watchObservedRunningTime="2025-10-03 13:35:03.581726873 +0000 UTC m=+272.785070290" Oct 03 13:35:04 crc kubenswrapper[4959]: I1003 13:35:04.938476 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:35:04 crc kubenswrapper[4959]: I1003 13:35:04.938629 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:35:04 crc kubenswrapper[4959]: I1003 13:35:04.991323 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:35:05 crc kubenswrapper[4959]: I1003 13:35:05.615182 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9h98" Oct 03 13:35:05 crc kubenswrapper[4959]: I1003 13:35:05.932280 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:35:05 crc kubenswrapper[4959]: I1003 13:35:05.932343 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:35:05 crc kubenswrapper[4959]: I1003 13:35:05.971099 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:35:06 crc kubenswrapper[4959]: I1003 13:35:06.624722 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8b59q" Oct 03 13:35:07 crc kubenswrapper[4959]: I1003 13:35:07.348881 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:35:07 crc kubenswrapper[4959]: I1003 13:35:07.348939 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:35:08 crc kubenswrapper[4959]: I1003 13:35:08.330807 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:35:08 crc kubenswrapper[4959]: I1003 13:35:08.332340 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:35:08 crc kubenswrapper[4959]: I1003 13:35:08.371317 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:35:08 crc kubenswrapper[4959]: I1003 13:35:08.399960 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5plkc" podUID="015eb8ae-9d78-41ed-b557-7cc23c222968" containerName="registry-server" probeResult="failure" output=< Oct 03 13:35:08 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 13:35:08 crc kubenswrapper[4959]: > Oct 03 13:35:08 crc kubenswrapper[4959]: I1003 13:35:08.626664 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4d9lm" Oct 03 13:35:17 crc kubenswrapper[4959]: I1003 13:35:17.388872 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:35:17 crc kubenswrapper[4959]: I1003 13:35:17.427453 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5plkc" Oct 03 13:36:36 crc kubenswrapper[4959]: I1003 13:36:36.044801 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:36:36 crc kubenswrapper[4959]: I1003 13:36:36.045580 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:37:06 crc kubenswrapper[4959]: I1003 13:37:06.047764 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:37:06 crc kubenswrapper[4959]: I1003 13:37:06.048345 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.044158 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.044792 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.044843 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.045501 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.045571 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540" gracePeriod=600 Oct 03 13:37:36 crc kubenswrapper[4959]: E1003 13:37:36.171770 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6639a7dd_61b2_4eb7_bf37_2ccbe2487dde.slice/crio-conmon-48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.394397 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540" exitCode=0 Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.394501 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540"} Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.394714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6"} Oct 03 13:37:36 crc kubenswrapper[4959]: I1003 13:37:36.394736 4959 scope.go:117] "RemoveContainer" containerID="78152f53edb20b4a385c2e67aaa2abd3f0d9103407b1535139f5f6540040351c" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.445357 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-95xsw"] Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.446503 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.456465 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-95xsw"] Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.553755 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-bound-sa-token\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.553821 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/057d1e97-1281-4017-8e21-538489d87d1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.553850 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/057d1e97-1281-4017-8e21-538489d87d1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.554042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-trusted-ca\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.554170 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.554263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-registry-certificates\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.554287 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-registry-tls\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.554310 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kdzf\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-kube-api-access-2kdzf\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.574751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.655767 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-registry-certificates\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.655971 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-registry-tls\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.655990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kdzf\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-kube-api-access-2kdzf\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.656011 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-bound-sa-token\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.656044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/057d1e97-1281-4017-8e21-538489d87d1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.656067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/057d1e97-1281-4017-8e21-538489d87d1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.656082 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-trusted-ca\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.656728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/057d1e97-1281-4017-8e21-538489d87d1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.657149 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-registry-certificates\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.657247 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/057d1e97-1281-4017-8e21-538489d87d1b-trusted-ca\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.661393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/057d1e97-1281-4017-8e21-538489d87d1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.661695 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-registry-tls\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.672064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-bound-sa-token\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.687939 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kdzf\" (UniqueName: \"kubernetes.io/projected/057d1e97-1281-4017-8e21-538489d87d1b-kube-api-access-2kdzf\") pod \"image-registry-66df7c8f76-95xsw\" (UID: \"057d1e97-1281-4017-8e21-538489d87d1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.763856 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:41 crc kubenswrapper[4959]: I1003 13:38:41.937404 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-95xsw"] Oct 03 13:38:42 crc kubenswrapper[4959]: I1003 13:38:42.730850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" event={"ID":"057d1e97-1281-4017-8e21-538489d87d1b","Type":"ContainerStarted","Data":"61cc18dad6988de4fd5a456979c8c6b58d34d7d5e76d5c239cbd2149b6ed2d3b"} Oct 03 13:38:42 crc kubenswrapper[4959]: I1003 13:38:42.730905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" event={"ID":"057d1e97-1281-4017-8e21-538489d87d1b","Type":"ContainerStarted","Data":"34adc2757d1146ce9340e3631e4af31728f6ff5c15a62c243c9e43cf996e621c"} Oct 03 13:38:42 crc kubenswrapper[4959]: I1003 13:38:42.731030 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:38:42 crc kubenswrapper[4959]: I1003 13:38:42.748573 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" podStartSLOduration=1.748555574 podStartE2EDuration="1.748555574s" podCreationTimestamp="2025-10-03 13:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:38:42.746573154 +0000 UTC m=+491.949916591" watchObservedRunningTime="2025-10-03 13:38:42.748555574 +0000 UTC m=+491.951898981" Oct 03 13:39:01 crc kubenswrapper[4959]: I1003 13:39:01.770029 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-95xsw" Oct 03 13:39:01 crc kubenswrapper[4959]: I1003 13:39:01.823441 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:39:26 crc kubenswrapper[4959]: I1003 13:39:26.862758 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" podUID="9f722c0f-dd0f-4d39-a444-687158d69894" containerName="registry" containerID="cri-o://10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7" gracePeriod=30 Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.185749 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.264862 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265007 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265043 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265089 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265112 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265842 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.265934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.266005 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4qfh\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.266391 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls\") pod \"9f722c0f-dd0f-4d39-a444-687158d69894\" (UID: \"9f722c0f-dd0f-4d39-a444-687158d69894\") " Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.266816 4959 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.267446 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.271129 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.271681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.272695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh" (OuterVolumeSpecName: "kube-api-access-c4qfh") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "kube-api-access-c4qfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.276766 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.276829 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.283376 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9f722c0f-dd0f-4d39-a444-687158d69894" (UID: "9f722c0f-dd0f-4d39-a444-687158d69894"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368266 4959 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368305 4959 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f722c0f-dd0f-4d39-a444-687158d69894-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368324 4959 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368333 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f722c0f-dd0f-4d39-a444-687158d69894-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368341 4959 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f722c0f-dd0f-4d39-a444-687158d69894-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.368350 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4qfh\" (UniqueName: \"kubernetes.io/projected/9f722c0f-dd0f-4d39-a444-687158d69894-kube-api-access-c4qfh\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.974829 4959 generic.go:334] "Generic (PLEG): container finished" podID="9f722c0f-dd0f-4d39-a444-687158d69894" containerID="10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7" exitCode=0 Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.974872 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.974879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" event={"ID":"9f722c0f-dd0f-4d39-a444-687158d69894","Type":"ContainerDied","Data":"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7"} Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.974915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-56clw" event={"ID":"9f722c0f-dd0f-4d39-a444-687158d69894","Type":"ContainerDied","Data":"e70712fadc1104751d075d9dbced34d48692e028839d8584e60b21803d510be2"} Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.974937 4959 scope.go:117] "RemoveContainer" containerID="10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.991221 4959 scope.go:117] "RemoveContainer" containerID="10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7" Oct 03 13:39:27 crc kubenswrapper[4959]: E1003 13:39:27.993352 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7\": container with ID starting with 10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7 not found: ID does not exist" containerID="10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.993429 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7"} err="failed to get container status \"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7\": rpc error: code = NotFound desc = could not find container \"10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7\": container with ID starting with 10b32273229e2b62f2c6a089bb51003c02b84f39006bbccc4176703d71a435a7 not found: ID does not exist" Oct 03 13:39:27 crc kubenswrapper[4959]: I1003 13:39:27.999071 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:39:28 crc kubenswrapper[4959]: I1003 13:39:28.001836 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-56clw"] Oct 03 13:39:29 crc kubenswrapper[4959]: I1003 13:39:29.692348 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f722c0f-dd0f-4d39-a444-687158d69894" path="/var/lib/kubelet/pods/9f722c0f-dd0f-4d39-a444-687158d69894/volumes" Oct 03 13:39:36 crc kubenswrapper[4959]: I1003 13:39:36.044171 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:39:36 crc kubenswrapper[4959]: I1003 13:39:36.045328 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:40:06 crc kubenswrapper[4959]: I1003 13:40:06.044759 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:40:06 crc kubenswrapper[4959]: I1003 13:40:06.045814 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.044323 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.044902 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.044985 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.045633 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.045762 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6" gracePeriod=600 Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.333988 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6" exitCode=0 Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.334049 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6"} Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.334337 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810"} Oct 03 13:40:36 crc kubenswrapper[4959]: I1003 13:40:36.334358 4959 scope.go:117] "RemoveContainer" containerID="48a65159914784e9d4673cec083bb53990fadcb93826bf6f4fb82026dab96540" Oct 03 13:42:36 crc kubenswrapper[4959]: I1003 13:42:36.044508 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:42:36 crc kubenswrapper[4959]: I1003 13:42:36.045629 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.500167 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.500959 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerName="controller-manager" containerID="cri-o://baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540" gracePeriod=30 Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.610550 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.610964 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerName="route-controller-manager" containerID="cri-o://957e53ca5689590cbe482d040611f2f1326aa04d89d60ac91af36ada86e6a393" gracePeriod=30 Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.841619 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.932362 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpxrr\" (UniqueName: \"kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr\") pod \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.932486 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert\") pod \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.932511 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config\") pod \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.932542 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca\") pod \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.932576 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles\") pod \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\" (UID: \"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b\") " Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933603 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca" (OuterVolumeSpecName: "client-ca") pod "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" (UID: "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933641 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" (UID: "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933719 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config" (OuterVolumeSpecName: "config") pod "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" (UID: "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933961 4959 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933979 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.933987 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.937784 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" (UID: "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:42:57 crc kubenswrapper[4959]: I1003 13:42:57.938029 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr" (OuterVolumeSpecName: "kube-api-access-gpxrr") pod "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" (UID: "5c43eaa0-47e1-4b1b-8b85-b5a278bb893b"). InnerVolumeSpecName "kube-api-access-gpxrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.034994 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpxrr\" (UniqueName: \"kubernetes.io/projected/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-kube-api-access-gpxrr\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.035044 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.039418 4959 generic.go:334] "Generic (PLEG): container finished" podID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerID="957e53ca5689590cbe482d040611f2f1326aa04d89d60ac91af36ada86e6a393" exitCode=0 Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.039457 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.039471 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" event={"ID":"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac","Type":"ContainerDied","Data":"957e53ca5689590cbe482d040611f2f1326aa04d89d60ac91af36ada86e6a393"} Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.039499 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" event={"ID":"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac","Type":"ContainerDied","Data":"aad9367e69cb42da54e93991aa974bc23fc61905c47fb3e30b7f9e298561dcae"} Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.039518 4959 scope.go:117] "RemoveContainer" containerID="957e53ca5689590cbe482d040611f2f1326aa04d89d60ac91af36ada86e6a393" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.041236 4959 generic.go:334] "Generic (PLEG): container finished" podID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerID="baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540" exitCode=0 Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.041259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" event={"ID":"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b","Type":"ContainerDied","Data":"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540"} Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.041275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" event={"ID":"5c43eaa0-47e1-4b1b-8b85-b5a278bb893b","Type":"ContainerDied","Data":"720ed1c7f3f0e1d8c22f04da443461f7bd89dc88f4a853c876e2b944ff12505e"} Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.041316 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-b5jrs" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.057476 4959 scope.go:117] "RemoveContainer" containerID="baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.070831 4959 scope.go:117] "RemoveContainer" containerID="baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540" Oct 03 13:42:58 crc kubenswrapper[4959]: E1003 13:42:58.073330 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540\": container with ID starting with baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540 not found: ID does not exist" containerID="baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.073385 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540"} err="failed to get container status \"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540\": rpc error: code = NotFound desc = could not find container \"baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540\": container with ID starting with baa190bba9a652abb292468665db2d65989d07de0fc3e2381fb60b213d091540 not found: ID does not exist" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.075282 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.078432 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-b5jrs"] Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.135487 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert\") pod \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.135626 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config\") pod \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.135654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca\") pod \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.135730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6k7d\" (UniqueName: \"kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d\") pod \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\" (UID: \"06dfaedd-2184-4d00-a0e6-98fa1fdc17ac\") " Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.136537 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config" (OuterVolumeSpecName: "config") pod "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" (UID: "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.136561 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca" (OuterVolumeSpecName: "client-ca") pod "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" (UID: "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.139287 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d" (OuterVolumeSpecName: "kube-api-access-h6k7d") pod "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" (UID: "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac"). InnerVolumeSpecName "kube-api-access-h6k7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.139439 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" (UID: "06dfaedd-2184-4d00-a0e6-98fa1fdc17ac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.237020 4959 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.237062 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.237071 4959 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:58 crc kubenswrapper[4959]: I1003 13:42:58.237080 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6k7d\" (UniqueName: \"kubernetes.io/projected/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac-kube-api-access-h6k7d\") on node \"crc\" DevicePath \"\"" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.049368 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.077297 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.081224 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-9zvr5"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.276876 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69dcc9c759-f6l5p"] Oct 03 13:42:59 crc kubenswrapper[4959]: E1003 13:42:59.277121 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerName="route-controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277133 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerName="route-controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: E1003 13:42:59.277159 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f722c0f-dd0f-4d39-a444-687158d69894" containerName="registry" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277168 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f722c0f-dd0f-4d39-a444-687158d69894" containerName="registry" Oct 03 13:42:59 crc kubenswrapper[4959]: E1003 13:42:59.277180 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerName="controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277205 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerName="controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277292 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" containerName="controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277301 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" containerName="route-controller-manager" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277307 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f722c0f-dd0f-4d39-a444-687158d69894" containerName="registry" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.277669 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.280624 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.281380 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.281444 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.281496 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.283595 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.283998 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284148 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284173 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284302 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284333 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284336 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284453 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284638 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.284912 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.290908 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.295898 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69dcc9c759-f6l5p"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.301558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352360 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btc72\" (UniqueName: \"kubernetes.io/projected/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-kube-api-access-btc72\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352419 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/666048c4-06e0-4f4b-9705-379c4b8fe702-serving-cert\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352439 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-proxy-ca-bundles\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352460 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-config\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352478 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-client-ca\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-serving-cert\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhvvc\" (UniqueName: \"kubernetes.io/projected/666048c4-06e0-4f4b-9705-379c4b8fe702-kube-api-access-bhvvc\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352542 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-client-ca\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.352635 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-config\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454051 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhvvc\" (UniqueName: \"kubernetes.io/projected/666048c4-06e0-4f4b-9705-379c4b8fe702-kube-api-access-bhvvc\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454126 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-client-ca\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454153 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-config\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454221 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btc72\" (UniqueName: \"kubernetes.io/projected/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-kube-api-access-btc72\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/666048c4-06e0-4f4b-9705-379c4b8fe702-serving-cert\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454286 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-proxy-ca-bundles\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-config\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454335 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-serving-cert\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.454354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-client-ca\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.455260 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-client-ca\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.455268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-client-ca\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.455765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-proxy-ca-bundles\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.455812 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/666048c4-06e0-4f4b-9705-379c4b8fe702-config\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.456240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-config\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.460076 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/666048c4-06e0-4f4b-9705-379c4b8fe702-serving-cert\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.460104 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-serving-cert\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.471637 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btc72\" (UniqueName: \"kubernetes.io/projected/47d1ee4a-5992-4361-b7e0-eb6544f78ee5-kube-api-access-btc72\") pod \"route-controller-manager-547c79c78f-cs89j\" (UID: \"47d1ee4a-5992-4361-b7e0-eb6544f78ee5\") " pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.478436 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhvvc\" (UniqueName: \"kubernetes.io/projected/666048c4-06e0-4f4b-9705-379c4b8fe702-kube-api-access-bhvvc\") pod \"controller-manager-69dcc9c759-f6l5p\" (UID: \"666048c4-06e0-4f4b-9705-379c4b8fe702\") " pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.596397 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.605012 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.708221 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06dfaedd-2184-4d00-a0e6-98fa1fdc17ac" path="/var/lib/kubelet/pods/06dfaedd-2184-4d00-a0e6-98fa1fdc17ac/volumes" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.709806 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c43eaa0-47e1-4b1b-8b85-b5a278bb893b" path="/var/lib/kubelet/pods/5c43eaa0-47e1-4b1b-8b85-b5a278bb893b/volumes" Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.807575 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j"] Oct 03 13:42:59 crc kubenswrapper[4959]: I1003 13:42:59.886498 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69dcc9c759-f6l5p"] Oct 03 13:42:59 crc kubenswrapper[4959]: W1003 13:42:59.895952 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod666048c4_06e0_4f4b_9705_379c4b8fe702.slice/crio-7180938d696fec8a5784bad37c3462a8c032b79daf82c78e6a6b837ec9783125 WatchSource:0}: Error finding container 7180938d696fec8a5784bad37c3462a8c032b79daf82c78e6a6b837ec9783125: Status 404 returned error can't find the container with id 7180938d696fec8a5784bad37c3462a8c032b79daf82c78e6a6b837ec9783125 Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.054501 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" event={"ID":"666048c4-06e0-4f4b-9705-379c4b8fe702","Type":"ContainerStarted","Data":"3c8b5c4b613dfcc2bf70cd728c13fd60ac1cba1f466ba6b001ba1a00363d70f3"} Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.054836 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" event={"ID":"666048c4-06e0-4f4b-9705-379c4b8fe702","Type":"ContainerStarted","Data":"7180938d696fec8a5784bad37c3462a8c032b79daf82c78e6a6b837ec9783125"} Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.055812 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.057878 4959 patch_prober.go:28] interesting pod/controller-manager-69dcc9c759-f6l5p container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.057909 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" podUID="666048c4-06e0-4f4b-9705-379c4b8fe702" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.059596 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" event={"ID":"47d1ee4a-5992-4361-b7e0-eb6544f78ee5","Type":"ContainerStarted","Data":"d5a60886dfea2660174e5efd1e7e62575bde6f24d95f958c8b0527c318ac70d9"} Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.059628 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" event={"ID":"47d1ee4a-5992-4361-b7e0-eb6544f78ee5","Type":"ContainerStarted","Data":"731af60416fc7cfcdfb72d266f771c1b954543fc5310a6a150331881dd5eeb9d"} Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.060277 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.061547 4959 patch_prober.go:28] interesting pod/route-controller-manager-547c79c78f-cs89j container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" start-of-body= Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.061589 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" podUID="47d1ee4a-5992-4361-b7e0-eb6544f78ee5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": dial tcp 10.217.0.63:8443: connect: connection refused" Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.077875 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" podStartSLOduration=3.077859368 podStartE2EDuration="3.077859368s" podCreationTimestamp="2025-10-03 13:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:43:00.075444448 +0000 UTC m=+749.278787865" watchObservedRunningTime="2025-10-03 13:43:00.077859368 +0000 UTC m=+749.281202785" Oct 03 13:43:00 crc kubenswrapper[4959]: I1003 13:43:00.094798 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" podStartSLOduration=3.094775682 podStartE2EDuration="3.094775682s" podCreationTimestamp="2025-10-03 13:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:43:00.093315556 +0000 UTC m=+749.296658973" watchObservedRunningTime="2025-10-03 13:43:00.094775682 +0000 UTC m=+749.298119099" Oct 03 13:43:01 crc kubenswrapper[4959]: I1003 13:43:01.067861 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-547c79c78f-cs89j" Oct 03 13:43:01 crc kubenswrapper[4959]: I1003 13:43:01.069217 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69dcc9c759-f6l5p" Oct 03 13:43:06 crc kubenswrapper[4959]: I1003 13:43:06.044378 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:43:06 crc kubenswrapper[4959]: I1003 13:43:06.044958 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:43:06 crc kubenswrapper[4959]: I1003 13:43:06.724161 4959 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.044697 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.045327 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.045379 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.046048 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.046131 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810" gracePeriod=600 Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.243894 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810" exitCode=0 Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.243939 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810"} Oct 03 13:43:36 crc kubenswrapper[4959]: I1003 13:43:36.243969 4959 scope.go:117] "RemoveContainer" containerID="0db3f11ceb8bb69b9ade0ae6e0ed7ad96db769dbdba74caf7fbebbc29bda9cc6" Oct 03 13:43:37 crc kubenswrapper[4959]: I1003 13:43:37.251177 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a"} Oct 03 13:43:55 crc kubenswrapper[4959]: I1003 13:43:55.853117 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:43:55 crc kubenswrapper[4959]: I1003 13:43:55.855084 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:55 crc kubenswrapper[4959]: I1003 13:43:55.869102 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.003924 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.004003 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw5dk\" (UniqueName: \"kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.004093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.105046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.105110 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw5dk\" (UniqueName: \"kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.105144 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.105516 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.105612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.123847 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw5dk\" (UniqueName: \"kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk\") pod \"certified-operators-8fjh7\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.174645 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:43:56 crc kubenswrapper[4959]: I1003 13:43:56.626345 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:43:57 crc kubenswrapper[4959]: I1003 13:43:57.371235 4959 generic.go:334] "Generic (PLEG): container finished" podID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerID="049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035" exitCode=0 Oct 03 13:43:57 crc kubenswrapper[4959]: I1003 13:43:57.371352 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerDied","Data":"049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035"} Oct 03 13:43:57 crc kubenswrapper[4959]: I1003 13:43:57.371503 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerStarted","Data":"3771c2e8a7a32de06222ab7ddb4a17a94493b9debd0e8c927237cba5f2d2e417"} Oct 03 13:43:57 crc kubenswrapper[4959]: I1003 13:43:57.373712 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:43:59 crc kubenswrapper[4959]: I1003 13:43:59.385984 4959 generic.go:334] "Generic (PLEG): container finished" podID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerID="c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2" exitCode=0 Oct 03 13:43:59 crc kubenswrapper[4959]: I1003 13:43:59.386054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerDied","Data":"c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2"} Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.393894 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerStarted","Data":"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3"} Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.421108 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8fjh7" podStartSLOduration=3.011365779 podStartE2EDuration="5.421086009s" podCreationTimestamp="2025-10-03 13:43:55 +0000 UTC" firstStartedPulling="2025-10-03 13:43:57.373509314 +0000 UTC m=+806.576852731" lastFinishedPulling="2025-10-03 13:43:59.783229504 +0000 UTC m=+808.986572961" observedRunningTime="2025-10-03 13:44:00.415642943 +0000 UTC m=+809.618986370" watchObservedRunningTime="2025-10-03 13:44:00.421086009 +0000 UTC m=+809.624429426" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.441951 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.443127 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.493508 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.567353 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.567396 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pcvq\" (UniqueName: \"kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.567416 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.668532 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.668614 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pcvq\" (UniqueName: \"kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.668649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.669641 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.669903 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.690064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pcvq\" (UniqueName: \"kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq\") pod \"community-operators-klnsf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:00 crc kubenswrapper[4959]: I1003 13:44:00.760769 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:01 crc kubenswrapper[4959]: I1003 13:44:01.217516 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:01 crc kubenswrapper[4959]: I1003 13:44:01.398930 4959 generic.go:334] "Generic (PLEG): container finished" podID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerID="896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb" exitCode=0 Oct 03 13:44:01 crc kubenswrapper[4959]: I1003 13:44:01.399034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerDied","Data":"896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb"} Oct 03 13:44:01 crc kubenswrapper[4959]: I1003 13:44:01.399080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerStarted","Data":"88f3f2f44b4df50bbf94bb49db7df98f3178dd69ac8569ce9b6512fc8d9ad0aa"} Oct 03 13:44:02 crc kubenswrapper[4959]: I1003 13:44:02.405490 4959 generic.go:334] "Generic (PLEG): container finished" podID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerID="1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503" exitCode=0 Oct 03 13:44:02 crc kubenswrapper[4959]: I1003 13:44:02.405742 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerDied","Data":"1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503"} Oct 03 13:44:04 crc kubenswrapper[4959]: I1003 13:44:04.422633 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerStarted","Data":"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e"} Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.176443 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.176504 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.215355 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.239778 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-klnsf" podStartSLOduration=4.294364129 podStartE2EDuration="6.239750214s" podCreationTimestamp="2025-10-03 13:44:00 +0000 UTC" firstStartedPulling="2025-10-03 13:44:01.40066419 +0000 UTC m=+810.604007597" lastFinishedPulling="2025-10-03 13:44:03.346050265 +0000 UTC m=+812.549393682" observedRunningTime="2025-10-03 13:44:04.449640411 +0000 UTC m=+813.652983848" watchObservedRunningTime="2025-10-03 13:44:06.239750214 +0000 UTC m=+815.443093661" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.460566 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.465399 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.468613 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.496523 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.658358 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7sl7\" (UniqueName: \"kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.658832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.658917 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.760208 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.760264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.760306 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7sl7\" (UniqueName: \"kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.760810 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.760946 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.785146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7sl7\" (UniqueName: \"kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7\") pod \"redhat-marketplace-4j4tb\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:06 crc kubenswrapper[4959]: I1003 13:44:06.803601 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:07 crc kubenswrapper[4959]: I1003 13:44:07.196036 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:07 crc kubenswrapper[4959]: W1003 13:44:07.205404 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc984ebac_423e_4a53_8470_323b28d1722d.slice/crio-e965c4fb3ed7004fc84028c15b630e53dea4d729eb538cb646783dbcde0e252d WatchSource:0}: Error finding container e965c4fb3ed7004fc84028c15b630e53dea4d729eb538cb646783dbcde0e252d: Status 404 returned error can't find the container with id e965c4fb3ed7004fc84028c15b630e53dea4d729eb538cb646783dbcde0e252d Oct 03 13:44:07 crc kubenswrapper[4959]: I1003 13:44:07.441774 4959 generic.go:334] "Generic (PLEG): container finished" podID="c984ebac-423e-4a53-8470-323b28d1722d" containerID="7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463" exitCode=0 Oct 03 13:44:07 crc kubenswrapper[4959]: I1003 13:44:07.442094 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerDied","Data":"7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463"} Oct 03 13:44:07 crc kubenswrapper[4959]: I1003 13:44:07.442184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerStarted","Data":"e965c4fb3ed7004fc84028c15b630e53dea4d729eb538cb646783dbcde0e252d"} Oct 03 13:44:08 crc kubenswrapper[4959]: I1003 13:44:08.448638 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerStarted","Data":"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9"} Oct 03 13:44:08 crc kubenswrapper[4959]: I1003 13:44:08.834773 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:44:08 crc kubenswrapper[4959]: I1003 13:44:08.835026 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8fjh7" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="registry-server" containerID="cri-o://1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3" gracePeriod=2 Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.215484 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.393233 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw5dk\" (UniqueName: \"kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk\") pod \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.393363 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content\") pod \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.393388 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities\") pod \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\" (UID: \"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4\") " Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.394698 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities" (OuterVolumeSpecName: "utilities") pod "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" (UID: "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.398499 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk" (OuterVolumeSpecName: "kube-api-access-bw5dk") pod "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" (UID: "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4"). InnerVolumeSpecName "kube-api-access-bw5dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.448274 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" (UID: "6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.456414 4959 generic.go:334] "Generic (PLEG): container finished" podID="c984ebac-423e-4a53-8470-323b28d1722d" containerID="5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9" exitCode=0 Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.456470 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerDied","Data":"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9"} Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.459349 4959 generic.go:334] "Generic (PLEG): container finished" podID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerID="1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3" exitCode=0 Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.459404 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fjh7" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.459402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerDied","Data":"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3"} Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.459461 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fjh7" event={"ID":"6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4","Type":"ContainerDied","Data":"3771c2e8a7a32de06222ab7ddb4a17a94493b9debd0e8c927237cba5f2d2e417"} Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.459488 4959 scope.go:117] "RemoveContainer" containerID="1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.479108 4959 scope.go:117] "RemoveContainer" containerID="c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.490418 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.494214 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.494244 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw5dk\" (UniqueName: \"kubernetes.io/projected/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-kube-api-access-bw5dk\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.494252 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.495390 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8fjh7"] Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.504835 4959 scope.go:117] "RemoveContainer" containerID="049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.518849 4959 scope.go:117] "RemoveContainer" containerID="1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3" Oct 03 13:44:09 crc kubenswrapper[4959]: E1003 13:44:09.519572 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3\": container with ID starting with 1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3 not found: ID does not exist" containerID="1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.519941 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3"} err="failed to get container status \"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3\": rpc error: code = NotFound desc = could not find container \"1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3\": container with ID starting with 1b86ad958338319d65e991a0121d68376158e6e3d67c4d1f73da189fa84140c3 not found: ID does not exist" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.519977 4959 scope.go:117] "RemoveContainer" containerID="c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2" Oct 03 13:44:09 crc kubenswrapper[4959]: E1003 13:44:09.520602 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2\": container with ID starting with c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2 not found: ID does not exist" containerID="c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.520657 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2"} err="failed to get container status \"c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2\": rpc error: code = NotFound desc = could not find container \"c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2\": container with ID starting with c89c875b024d63e144ce86824880c8aa7c159096a89726e859d9d001c79e50b2 not found: ID does not exist" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.520692 4959 scope.go:117] "RemoveContainer" containerID="049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035" Oct 03 13:44:09 crc kubenswrapper[4959]: E1003 13:44:09.521020 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035\": container with ID starting with 049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035 not found: ID does not exist" containerID="049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.521061 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035"} err="failed to get container status \"049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035\": rpc error: code = NotFound desc = could not find container \"049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035\": container with ID starting with 049586e4533e31ad27fdc386803e868e6889d81f5ca65b4b46c61de98bdba035 not found: ID does not exist" Oct 03 13:44:09 crc kubenswrapper[4959]: I1003 13:44:09.696687 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" path="/var/lib/kubelet/pods/6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4/volumes" Oct 03 13:44:10 crc kubenswrapper[4959]: I1003 13:44:10.465953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerStarted","Data":"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46"} Oct 03 13:44:10 crc kubenswrapper[4959]: I1003 13:44:10.761714 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:10 crc kubenswrapper[4959]: I1003 13:44:10.761983 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:10 crc kubenswrapper[4959]: I1003 13:44:10.802225 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:10 crc kubenswrapper[4959]: I1003 13:44:10.819098 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4j4tb" podStartSLOduration=2.393046683 podStartE2EDuration="4.819081054s" podCreationTimestamp="2025-10-03 13:44:06 +0000 UTC" firstStartedPulling="2025-10-03 13:44:07.443497337 +0000 UTC m=+816.646840744" lastFinishedPulling="2025-10-03 13:44:09.869531658 +0000 UTC m=+819.072875115" observedRunningTime="2025-10-03 13:44:10.484627611 +0000 UTC m=+819.687971018" watchObservedRunningTime="2025-10-03 13:44:10.819081054 +0000 UTC m=+820.022424471" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.241500 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:11 crc kubenswrapper[4959]: E1003 13:44:11.241738 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="extract-utilities" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.241754 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="extract-utilities" Oct 03 13:44:11 crc kubenswrapper[4959]: E1003 13:44:11.241764 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="extract-content" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.241771 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="extract-content" Oct 03 13:44:11 crc kubenswrapper[4959]: E1003 13:44:11.241782 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="registry-server" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.241791 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="registry-server" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.241910 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9774f3-0eb0-4f99-bdcd-7a9af7cbf7e4" containerName="registry-server" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.246121 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.253859 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.418799 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.418850 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.419117 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d67mj\" (UniqueName: \"kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.504019 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.520806 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.520876 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.520960 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d67mj\" (UniqueName: \"kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.521336 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.521447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.544269 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d67mj\" (UniqueName: \"kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj\") pod \"redhat-operators-djnhm\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.566645 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:11 crc kubenswrapper[4959]: I1003 13:44:11.751011 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:11 crc kubenswrapper[4959]: W1003 13:44:11.757488 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fd40a96_aca0_4ee8_a56e_f9e811d94206.slice/crio-f31a0e4e9c16805db1b0ed15a1ac7f536c823ea251dd0f590c6657c5a29d9833 WatchSource:0}: Error finding container f31a0e4e9c16805db1b0ed15a1ac7f536c823ea251dd0f590c6657c5a29d9833: Status 404 returned error can't find the container with id f31a0e4e9c16805db1b0ed15a1ac7f536c823ea251dd0f590c6657c5a29d9833 Oct 03 13:44:12 crc kubenswrapper[4959]: I1003 13:44:12.476240 4959 generic.go:334] "Generic (PLEG): container finished" podID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerID="3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4" exitCode=0 Oct 03 13:44:12 crc kubenswrapper[4959]: I1003 13:44:12.477338 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerDied","Data":"3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4"} Oct 03 13:44:12 crc kubenswrapper[4959]: I1003 13:44:12.477365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerStarted","Data":"f31a0e4e9c16805db1b0ed15a1ac7f536c823ea251dd0f590c6657c5a29d9833"} Oct 03 13:44:14 crc kubenswrapper[4959]: I1003 13:44:14.633154 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:14 crc kubenswrapper[4959]: I1003 13:44:14.633365 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-klnsf" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="registry-server" containerID="cri-o://73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e" gracePeriod=2 Oct 03 13:44:14 crc kubenswrapper[4959]: I1003 13:44:14.987246 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.163484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities\") pod \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.163578 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pcvq\" (UniqueName: \"kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq\") pod \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.163627 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content\") pod \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\" (UID: \"39bd7c92-f476-47ac-b3e3-988120c2a1cf\") " Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.164494 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities" (OuterVolumeSpecName: "utilities") pod "39bd7c92-f476-47ac-b3e3-988120c2a1cf" (UID: "39bd7c92-f476-47ac-b3e3-988120c2a1cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.168829 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq" (OuterVolumeSpecName: "kube-api-access-7pcvq") pod "39bd7c92-f476-47ac-b3e3-988120c2a1cf" (UID: "39bd7c92-f476-47ac-b3e3-988120c2a1cf"). InnerVolumeSpecName "kube-api-access-7pcvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.209257 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39bd7c92-f476-47ac-b3e3-988120c2a1cf" (UID: "39bd7c92-f476-47ac-b3e3-988120c2a1cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.266061 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pcvq\" (UniqueName: \"kubernetes.io/projected/39bd7c92-f476-47ac-b3e3-988120c2a1cf-kube-api-access-7pcvq\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.266611 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.266627 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39bd7c92-f476-47ac-b3e3-988120c2a1cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.493347 4959 generic.go:334] "Generic (PLEG): container finished" podID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerID="c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6" exitCode=0 Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.493460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerDied","Data":"c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6"} Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.505210 4959 generic.go:334] "Generic (PLEG): container finished" podID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerID="73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e" exitCode=0 Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.505258 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerDied","Data":"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e"} Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.505289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klnsf" event={"ID":"39bd7c92-f476-47ac-b3e3-988120c2a1cf","Type":"ContainerDied","Data":"88f3f2f44b4df50bbf94bb49db7df98f3178dd69ac8569ce9b6512fc8d9ad0aa"} Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.505311 4959 scope.go:117] "RemoveContainer" containerID="73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.505450 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klnsf" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.526704 4959 scope.go:117] "RemoveContainer" containerID="1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.534811 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.540151 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-klnsf"] Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.560985 4959 scope.go:117] "RemoveContainer" containerID="896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.572355 4959 scope.go:117] "RemoveContainer" containerID="73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e" Oct 03 13:44:15 crc kubenswrapper[4959]: E1003 13:44:15.572774 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e\": container with ID starting with 73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e not found: ID does not exist" containerID="73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.572815 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e"} err="failed to get container status \"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e\": rpc error: code = NotFound desc = could not find container \"73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e\": container with ID starting with 73dc30e950890fe409d8f4294092d87e6a35782f3012c7f8e98c23fb1f0d959e not found: ID does not exist" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.572841 4959 scope.go:117] "RemoveContainer" containerID="1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503" Oct 03 13:44:15 crc kubenswrapper[4959]: E1003 13:44:15.573090 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503\": container with ID starting with 1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503 not found: ID does not exist" containerID="1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.573115 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503"} err="failed to get container status \"1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503\": rpc error: code = NotFound desc = could not find container \"1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503\": container with ID starting with 1ca86d02a9d639a8e3f7754a218b250adba563f39da023636739e47d59e3f503 not found: ID does not exist" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.573131 4959 scope.go:117] "RemoveContainer" containerID="896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb" Oct 03 13:44:15 crc kubenswrapper[4959]: E1003 13:44:15.573482 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb\": container with ID starting with 896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb not found: ID does not exist" containerID="896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.573508 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb"} err="failed to get container status \"896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb\": rpc error: code = NotFound desc = could not find container \"896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb\": container with ID starting with 896847b81150295fe067b9f11917b8fbb7ab52c305beee03ab3754965fa46beb not found: ID does not exist" Oct 03 13:44:15 crc kubenswrapper[4959]: I1003 13:44:15.692600 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" path="/var/lib/kubelet/pods/39bd7c92-f476-47ac-b3e3-988120c2a1cf/volumes" Oct 03 13:44:16 crc kubenswrapper[4959]: I1003 13:44:16.512225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerStarted","Data":"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f"} Oct 03 13:44:16 crc kubenswrapper[4959]: I1003 13:44:16.527956 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djnhm" podStartSLOduration=1.8515149659999999 podStartE2EDuration="5.5279399s" podCreationTimestamp="2025-10-03 13:44:11 +0000 UTC" firstStartedPulling="2025-10-03 13:44:12.477864029 +0000 UTC m=+821.681207446" lastFinishedPulling="2025-10-03 13:44:16.154288963 +0000 UTC m=+825.357632380" observedRunningTime="2025-10-03 13:44:16.527011138 +0000 UTC m=+825.730354555" watchObservedRunningTime="2025-10-03 13:44:16.5279399 +0000 UTC m=+825.731283317" Oct 03 13:44:16 crc kubenswrapper[4959]: I1003 13:44:16.805262 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:16 crc kubenswrapper[4959]: I1003 13:44:16.805311 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:16 crc kubenswrapper[4959]: I1003 13:44:16.843257 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:17 crc kubenswrapper[4959]: I1003 13:44:17.553682 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:19 crc kubenswrapper[4959]: I1003 13:44:19.833715 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:19 crc kubenswrapper[4959]: I1003 13:44:19.834220 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4j4tb" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="registry-server" containerID="cri-o://468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46" gracePeriod=2 Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.180423 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.332588 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities\") pod \"c984ebac-423e-4a53-8470-323b28d1722d\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.332665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7sl7\" (UniqueName: \"kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7\") pod \"c984ebac-423e-4a53-8470-323b28d1722d\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.332691 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content\") pod \"c984ebac-423e-4a53-8470-323b28d1722d\" (UID: \"c984ebac-423e-4a53-8470-323b28d1722d\") " Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.333658 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities" (OuterVolumeSpecName: "utilities") pod "c984ebac-423e-4a53-8470-323b28d1722d" (UID: "c984ebac-423e-4a53-8470-323b28d1722d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.338504 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7" (OuterVolumeSpecName: "kube-api-access-g7sl7") pod "c984ebac-423e-4a53-8470-323b28d1722d" (UID: "c984ebac-423e-4a53-8470-323b28d1722d"). InnerVolumeSpecName "kube-api-access-g7sl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.344404 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c984ebac-423e-4a53-8470-323b28d1722d" (UID: "c984ebac-423e-4a53-8470-323b28d1722d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.433992 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.434275 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7sl7\" (UniqueName: \"kubernetes.io/projected/c984ebac-423e-4a53-8470-323b28d1722d-kube-api-access-g7sl7\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.434390 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c984ebac-423e-4a53-8470-323b28d1722d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.535486 4959 generic.go:334] "Generic (PLEG): container finished" podID="c984ebac-423e-4a53-8470-323b28d1722d" containerID="468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46" exitCode=0 Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.535534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerDied","Data":"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46"} Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.535561 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4j4tb" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.535595 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4j4tb" event={"ID":"c984ebac-423e-4a53-8470-323b28d1722d","Type":"ContainerDied","Data":"e965c4fb3ed7004fc84028c15b630e53dea4d729eb538cb646783dbcde0e252d"} Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.535611 4959 scope.go:117] "RemoveContainer" containerID="468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.563919 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.567498 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4j4tb"] Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.567920 4959 scope.go:117] "RemoveContainer" containerID="5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.592728 4959 scope.go:117] "RemoveContainer" containerID="7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.609533 4959 scope.go:117] "RemoveContainer" containerID="468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46" Oct 03 13:44:20 crc kubenswrapper[4959]: E1003 13:44:20.610111 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46\": container with ID starting with 468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46 not found: ID does not exist" containerID="468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.610258 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46"} err="failed to get container status \"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46\": rpc error: code = NotFound desc = could not find container \"468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46\": container with ID starting with 468e3851755ff0ee1e07c1e04133edc6e05f6c5eb1d66ae984d1e0af5ee64f46 not found: ID does not exist" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.610330 4959 scope.go:117] "RemoveContainer" containerID="5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9" Oct 03 13:44:20 crc kubenswrapper[4959]: E1003 13:44:20.610916 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9\": container with ID starting with 5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9 not found: ID does not exist" containerID="5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.611044 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9"} err="failed to get container status \"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9\": rpc error: code = NotFound desc = could not find container \"5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9\": container with ID starting with 5ecfb3bf5da9c403e0e783ae8b7dd0f520f036969fecda85577d5aa6147468e9 not found: ID does not exist" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.611147 4959 scope.go:117] "RemoveContainer" containerID="7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463" Oct 03 13:44:20 crc kubenswrapper[4959]: E1003 13:44:20.611674 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463\": container with ID starting with 7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463 not found: ID does not exist" containerID="7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463" Oct 03 13:44:20 crc kubenswrapper[4959]: I1003 13:44:20.611708 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463"} err="failed to get container status \"7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463\": rpc error: code = NotFound desc = could not find container \"7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463\": container with ID starting with 7302f485fea0d4e70488a3b3d3f788d8a685a7162e08de7a65741c3e554a5463 not found: ID does not exist" Oct 03 13:44:21 crc kubenswrapper[4959]: I1003 13:44:21.567500 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:21 crc kubenswrapper[4959]: I1003 13:44:21.567768 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:21 crc kubenswrapper[4959]: I1003 13:44:21.631065 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:21 crc kubenswrapper[4959]: I1003 13:44:21.691651 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c984ebac-423e-4a53-8470-323b28d1722d" path="/var/lib/kubelet/pods/c984ebac-423e-4a53-8470-323b28d1722d/volumes" Oct 03 13:44:22 crc kubenswrapper[4959]: I1003 13:44:22.591794 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:23 crc kubenswrapper[4959]: I1003 13:44:23.433772 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:24 crc kubenswrapper[4959]: I1003 13:44:24.555579 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-djnhm" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="registry-server" containerID="cri-o://a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f" gracePeriod=2 Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.429177 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.499541 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d67mj\" (UniqueName: \"kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj\") pod \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.499601 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities\") pod \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.499725 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content\") pod \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\" (UID: \"1fd40a96-aca0-4ee8-a56e-f9e811d94206\") " Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.501226 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities" (OuterVolumeSpecName: "utilities") pod "1fd40a96-aca0-4ee8-a56e-f9e811d94206" (UID: "1fd40a96-aca0-4ee8-a56e-f9e811d94206"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.505158 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj" (OuterVolumeSpecName: "kube-api-access-d67mj") pod "1fd40a96-aca0-4ee8-a56e-f9e811d94206" (UID: "1fd40a96-aca0-4ee8-a56e-f9e811d94206"). InnerVolumeSpecName "kube-api-access-d67mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.561713 4959 generic.go:334] "Generic (PLEG): container finished" podID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerID="a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f" exitCode=0 Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.561759 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerDied","Data":"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f"} Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.561781 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djnhm" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.561803 4959 scope.go:117] "RemoveContainer" containerID="a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.561788 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djnhm" event={"ID":"1fd40a96-aca0-4ee8-a56e-f9e811d94206","Type":"ContainerDied","Data":"f31a0e4e9c16805db1b0ed15a1ac7f536c823ea251dd0f590c6657c5a29d9833"} Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.577104 4959 scope.go:117] "RemoveContainer" containerID="c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.591379 4959 scope.go:117] "RemoveContainer" containerID="3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.601324 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.601349 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d67mj\" (UniqueName: \"kubernetes.io/projected/1fd40a96-aca0-4ee8-a56e-f9e811d94206-kube-api-access-d67mj\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.609246 4959 scope.go:117] "RemoveContainer" containerID="a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f" Oct 03 13:44:25 crc kubenswrapper[4959]: E1003 13:44:25.609639 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f\": container with ID starting with a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f not found: ID does not exist" containerID="a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.609674 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f"} err="failed to get container status \"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f\": rpc error: code = NotFound desc = could not find container \"a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f\": container with ID starting with a7228fd5175f358f15ff32fc08568315a609bd199c7e72682e94f16afcd9d16f not found: ID does not exist" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.609694 4959 scope.go:117] "RemoveContainer" containerID="c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6" Oct 03 13:44:25 crc kubenswrapper[4959]: E1003 13:44:25.610056 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6\": container with ID starting with c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6 not found: ID does not exist" containerID="c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.610108 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6"} err="failed to get container status \"c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6\": rpc error: code = NotFound desc = could not find container \"c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6\": container with ID starting with c2f66cbc50b5c03e10d1ac8833b455c20bbae7aef0770bca2d97f71b9f1032a6 not found: ID does not exist" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.610141 4959 scope.go:117] "RemoveContainer" containerID="3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4" Oct 03 13:44:25 crc kubenswrapper[4959]: E1003 13:44:25.610415 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4\": container with ID starting with 3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4 not found: ID does not exist" containerID="3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.610436 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4"} err="failed to get container status \"3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4\": rpc error: code = NotFound desc = could not find container \"3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4\": container with ID starting with 3fbef6595fcef97dd00d4d640db640fbcf046e18a740e19c02121666f109c8a4 not found: ID does not exist" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.625038 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fd40a96-aca0-4ee8-a56e-f9e811d94206" (UID: "1fd40a96-aca0-4ee8-a56e-f9e811d94206"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.702764 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fd40a96-aca0-4ee8-a56e-f9e811d94206-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.878742 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:25 crc kubenswrapper[4959]: I1003 13:44:25.884904 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-djnhm"] Oct 03 13:44:27 crc kubenswrapper[4959]: I1003 13:44:27.692609 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" path="/var/lib/kubelet/pods/1fd40a96-aca0-4ee8-a56e-f9e811d94206/volumes" Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.570684 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xqbk4"] Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665105 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-controller" containerID="cri-o://eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665124 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="nbdb" containerID="cri-o://023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665241 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="sbdb" containerID="cri-o://2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665258 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-node" containerID="cri-o://cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665286 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="northd" containerID="cri-o://303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665324 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-acl-logging" containerID="cri-o://c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.665242 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.702890 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" containerID="cri-o://1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" gracePeriod=30 Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.975740 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/3.log" Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.979031 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovn-acl-logging/0.log" Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.980243 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovn-controller/0.log" Oct 03 13:44:44 crc kubenswrapper[4959]: I1003 13:44:44.980651 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039061 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-srzq5"] Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039305 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039318 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039329 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kubecfg-setup" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039335 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kubecfg-setup" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039345 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039352 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039360 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039368 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039375 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039381 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039392 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039407 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039415 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-node" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039421 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-node" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039428 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039434 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039441 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039446 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039452 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039457 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039464 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-acl-logging" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039470 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-acl-logging" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039477 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039483 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039489 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039496 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039505 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="northd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039510 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="northd" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039518 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039524 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="extract-content" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039530 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039536 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039544 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039550 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039557 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="nbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039562 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="nbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039570 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039576 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="extract-utilities" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039582 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="sbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039587 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="sbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039668 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c984ebac-423e-4a53-8470-323b28d1722d" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039677 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039683 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039691 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="sbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039699 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bd7c92-f476-47ac-b3e3-988120c2a1cf" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039708 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039714 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="nbdb" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039721 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="northd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039728 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd40a96-aca0-4ee8-a56e-f9e811d94206" containerName="registry-server" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039735 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-acl-logging" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039742 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-node" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039749 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovn-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039756 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.039836 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039844 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039913 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.039923 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.040003 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.040010 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" containerName="ovnkube-controller" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.041307 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.132712 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.132821 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.132832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.132866 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133065 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash" (OuterVolumeSpecName: "host-slash") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133156 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133180 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133363 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tjvp\" (UniqueName: \"kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133419 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133446 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133879 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133950 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.133974 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134023 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log" (OuterVolumeSpecName: "node-log") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134052 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134074 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134106 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134121 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134168 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134185 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134149 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134479 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134504 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134522 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134549 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134636 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134677 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket" (OuterVolumeSpecName: "log-socket") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134658 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134729 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134752 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib\") pod \"3185a69a-b160-4257-afe1-a2165aa6e961\" (UID: \"3185a69a-b160-4257-afe1-a2165aa6e961\") " Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134780 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.134998 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135220 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135466 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-script-lib\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135523 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-netd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-config\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-log-socket\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135643 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-systemd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-ovn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135712 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-node-log\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135731 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-env-overrides\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135852 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-systemd-units\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135887 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-kubelet\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-bin\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135925 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401139a0-a614-4077-bbfd-e62293a96516-ovn-node-metrics-cert\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135949 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gjvn\" (UniqueName: \"kubernetes.io/projected/401139a0-a614-4077-bbfd-e62293a96516-kube-api-access-8gjvn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.135974 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136114 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-slash\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136154 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-netns\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136174 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-var-lib-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136286 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-etc-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136337 4959 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136348 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136360 4959 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136370 4959 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136381 4959 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-slash\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136389 4959 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136398 4959 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136407 4959 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136415 4959 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-node-log\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136424 4959 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136432 4959 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136441 4959 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136449 4959 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136459 4959 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136467 4959 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3185a69a-b160-4257-afe1-a2165aa6e961-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136475 4959 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-log-socket\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.136484 4959 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.138539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp" (OuterVolumeSpecName: "kube-api-access-8tjvp") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "kube-api-access-8tjvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.138783 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.146605 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3185a69a-b160-4257-afe1-a2165aa6e961" (UID: "3185a69a-b160-4257-afe1-a2165aa6e961"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.237839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-kubelet\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.237910 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-bin\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.237934 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-kubelet\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.237939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401139a0-a614-4077-bbfd-e62293a96516-ovn-node-metrics-cert\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gjvn\" (UniqueName: \"kubernetes.io/projected/401139a0-a614-4077-bbfd-e62293a96516-kube-api-access-8gjvn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238011 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-bin\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238050 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238031 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238217 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-slash\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238258 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-netns\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238281 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-var-lib-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238317 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-slash\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238319 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238356 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-ovn-kubernetes\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238356 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-run-netns\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238373 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-etc-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238356 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-etc-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238411 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-script-lib\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238465 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238471 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-var-lib-openvswitch\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238469 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-netd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238581 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-config\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238492 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-host-cni-netd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238641 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-log-socket\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238678 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-systemd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-systemd\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238884 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-ovn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238928 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-run-ovn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238945 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-node-log\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238980 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-env-overrides\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238995 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-node-log\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.238885 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-log-socket\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239069 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-systemd-units\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239180 4959 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3185a69a-b160-4257-afe1-a2165aa6e961-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-script-lib\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239256 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3185a69a-b160-4257-afe1-a2165aa6e961-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239297 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/401139a0-a614-4077-bbfd-e62293a96516-systemd-units\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239313 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tjvp\" (UniqueName: \"kubernetes.io/projected/3185a69a-b160-4257-afe1-a2165aa6e961-kube-api-access-8tjvp\") on node \"crc\" DevicePath \"\"" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239381 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-ovnkube-config\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.239486 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/401139a0-a614-4077-bbfd-e62293a96516-env-overrides\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.241744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/401139a0-a614-4077-bbfd-e62293a96516-ovn-node-metrics-cert\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.258768 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gjvn\" (UniqueName: \"kubernetes.io/projected/401139a0-a614-4077-bbfd-e62293a96516-kube-api-access-8gjvn\") pod \"ovnkube-node-srzq5\" (UID: \"401139a0-a614-4077-bbfd-e62293a96516\") " pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.353423 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:45 crc kubenswrapper[4959]: W1003 13:44:45.387641 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod401139a0_a614_4077_bbfd_e62293a96516.slice/crio-39d8f2cc40a124dfbae5efc07fb35874a5a670bf71e5a2dcff441b53a9555fff WatchSource:0}: Error finding container 39d8f2cc40a124dfbae5efc07fb35874a5a670bf71e5a2dcff441b53a9555fff: Status 404 returned error can't find the container with id 39d8f2cc40a124dfbae5efc07fb35874a5a670bf71e5a2dcff441b53a9555fff Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.672376 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/2.log" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.673233 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/1.log" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.673275 4959 generic.go:334] "Generic (PLEG): container finished" podID="117b68eb-14c1-456d-8c37-6259f7328587" containerID="d2f163aff2d8dfb8ab5462c2397a3b695da1e7a41334a2cb0dd2b7bb93530ea4" exitCode=2 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.673331 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerDied","Data":"d2f163aff2d8dfb8ab5462c2397a3b695da1e7a41334a2cb0dd2b7bb93530ea4"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.673364 4959 scope.go:117] "RemoveContainer" containerID="7fd5759d91c69a838bc236cfa61d0854feef402e518f8078c20db57d4cadcbe0" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.674607 4959 scope.go:117] "RemoveContainer" containerID="d2f163aff2d8dfb8ab5462c2397a3b695da1e7a41334a2cb0dd2b7bb93530ea4" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.675244 4959 generic.go:334] "Generic (PLEG): container finished" podID="401139a0-a614-4077-bbfd-e62293a96516" containerID="4418e6931d8357777165163be3a6072434afb711aed901ae4f13f7b494b8d507" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.675324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerDied","Data":"4418e6931d8357777165163be3a6072434afb711aed901ae4f13f7b494b8d507"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.675372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"39d8f2cc40a124dfbae5efc07fb35874a5a670bf71e5a2dcff441b53a9555fff"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.680040 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovnkube-controller/3.log" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.683520 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovn-acl-logging/0.log" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.684322 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xqbk4_3185a69a-b160-4257-afe1-a2165aa6e961/ovn-controller/0.log" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.685849 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.685915 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.685936 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.685956 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.685975 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.686051 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" exitCode=0 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.686075 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" exitCode=143 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.686573 4959 generic.go:334] "Generic (PLEG): container finished" podID="3185a69a-b160-4257-afe1-a2165aa6e961" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" exitCode=143 Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.686536 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699223 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699294 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699318 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699358 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699377 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699395 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699413 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699426 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699437 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699449 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699459 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699472 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699483 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699494 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699504 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699519 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699534 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699547 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699558 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699569 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699580 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699590 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699600 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699610 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699623 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699634 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699647 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699662 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699674 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699685 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699697 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699708 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699718 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699729 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699739 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699749 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699760 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699775 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xqbk4" event={"ID":"3185a69a-b160-4257-afe1-a2165aa6e961","Type":"ContainerDied","Data":"4869ec5f0a19823454df7cbb763a45e2ceae6d03714f2576a9cabe43b08a7648"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699791 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699802 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699813 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699827 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699838 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699849 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699862 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699873 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699884 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.699895 4959 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.707116 4959 scope.go:117] "RemoveContainer" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.733326 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.768677 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xqbk4"] Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.771133 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xqbk4"] Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.774545 4959 scope.go:117] "RemoveContainer" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.813938 4959 scope.go:117] "RemoveContainer" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.828682 4959 scope.go:117] "RemoveContainer" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.839766 4959 scope.go:117] "RemoveContainer" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.861583 4959 scope.go:117] "RemoveContainer" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.900094 4959 scope.go:117] "RemoveContainer" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.926380 4959 scope.go:117] "RemoveContainer" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.943263 4959 scope.go:117] "RemoveContainer" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.972494 4959 scope.go:117] "RemoveContainer" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.973725 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": container with ID starting with 1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546 not found: ID does not exist" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.973777 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} err="failed to get container status \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": rpc error: code = NotFound desc = could not find container \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": container with ID starting with 1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.973813 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.974139 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": container with ID starting with 35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1 not found: ID does not exist" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974168 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} err="failed to get container status \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": rpc error: code = NotFound desc = could not find container \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": container with ID starting with 35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974275 4959 scope.go:117] "RemoveContainer" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.974576 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": container with ID starting with 2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b not found: ID does not exist" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974599 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} err="failed to get container status \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": rpc error: code = NotFound desc = could not find container \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": container with ID starting with 2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974611 4959 scope.go:117] "RemoveContainer" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.974867 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": container with ID starting with 023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab not found: ID does not exist" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974899 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} err="failed to get container status \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": rpc error: code = NotFound desc = could not find container \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": container with ID starting with 023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.974922 4959 scope.go:117] "RemoveContainer" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.975185 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": container with ID starting with 303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15 not found: ID does not exist" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975219 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} err="failed to get container status \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": rpc error: code = NotFound desc = could not find container \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": container with ID starting with 303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975233 4959 scope.go:117] "RemoveContainer" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.975460 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": container with ID starting with c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2 not found: ID does not exist" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975502 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} err="failed to get container status \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": rpc error: code = NotFound desc = could not find container \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": container with ID starting with c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975518 4959 scope.go:117] "RemoveContainer" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.975755 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": container with ID starting with cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd not found: ID does not exist" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975779 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} err="failed to get container status \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": rpc error: code = NotFound desc = could not find container \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": container with ID starting with cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975792 4959 scope.go:117] "RemoveContainer" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.975977 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": container with ID starting with c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8 not found: ID does not exist" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.975996 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} err="failed to get container status \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": rpc error: code = NotFound desc = could not find container \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": container with ID starting with c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.976008 4959 scope.go:117] "RemoveContainer" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.976238 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": container with ID starting with eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47 not found: ID does not exist" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.976258 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} err="failed to get container status \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": rpc error: code = NotFound desc = could not find container \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": container with ID starting with eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.976272 4959 scope.go:117] "RemoveContainer" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: E1003 13:44:45.979026 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": container with ID starting with c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598 not found: ID does not exist" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979083 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} err="failed to get container status \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": rpc error: code = NotFound desc = could not find container \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": container with ID starting with c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979099 4959 scope.go:117] "RemoveContainer" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979326 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} err="failed to get container status \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": rpc error: code = NotFound desc = could not find container \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": container with ID starting with 1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979353 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979669 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} err="failed to get container status \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": rpc error: code = NotFound desc = could not find container \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": container with ID starting with 35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.979696 4959 scope.go:117] "RemoveContainer" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980153 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} err="failed to get container status \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": rpc error: code = NotFound desc = could not find container \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": container with ID starting with 2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980179 4959 scope.go:117] "RemoveContainer" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980406 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} err="failed to get container status \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": rpc error: code = NotFound desc = could not find container \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": container with ID starting with 023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980430 4959 scope.go:117] "RemoveContainer" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980630 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} err="failed to get container status \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": rpc error: code = NotFound desc = could not find container \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": container with ID starting with 303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980654 4959 scope.go:117] "RemoveContainer" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980869 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} err="failed to get container status \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": rpc error: code = NotFound desc = could not find container \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": container with ID starting with c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.980899 4959 scope.go:117] "RemoveContainer" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981117 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} err="failed to get container status \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": rpc error: code = NotFound desc = could not find container \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": container with ID starting with cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981143 4959 scope.go:117] "RemoveContainer" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981370 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} err="failed to get container status \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": rpc error: code = NotFound desc = could not find container \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": container with ID starting with c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981395 4959 scope.go:117] "RemoveContainer" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981602 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} err="failed to get container status \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": rpc error: code = NotFound desc = could not find container \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": container with ID starting with eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981625 4959 scope.go:117] "RemoveContainer" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981853 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} err="failed to get container status \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": rpc error: code = NotFound desc = could not find container \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": container with ID starting with c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.981879 4959 scope.go:117] "RemoveContainer" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982105 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} err="failed to get container status \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": rpc error: code = NotFound desc = could not find container \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": container with ID starting with 1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982127 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982389 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} err="failed to get container status \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": rpc error: code = NotFound desc = could not find container \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": container with ID starting with 35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982413 4959 scope.go:117] "RemoveContainer" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982613 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} err="failed to get container status \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": rpc error: code = NotFound desc = could not find container \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": container with ID starting with 2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982638 4959 scope.go:117] "RemoveContainer" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.982987 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} err="failed to get container status \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": rpc error: code = NotFound desc = could not find container \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": container with ID starting with 023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983011 4959 scope.go:117] "RemoveContainer" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983203 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} err="failed to get container status \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": rpc error: code = NotFound desc = could not find container \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": container with ID starting with 303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983224 4959 scope.go:117] "RemoveContainer" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983489 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} err="failed to get container status \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": rpc error: code = NotFound desc = could not find container \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": container with ID starting with c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983512 4959 scope.go:117] "RemoveContainer" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983720 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} err="failed to get container status \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": rpc error: code = NotFound desc = could not find container \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": container with ID starting with cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983739 4959 scope.go:117] "RemoveContainer" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983903 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} err="failed to get container status \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": rpc error: code = NotFound desc = could not find container \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": container with ID starting with c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.983927 4959 scope.go:117] "RemoveContainer" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984123 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} err="failed to get container status \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": rpc error: code = NotFound desc = could not find container \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": container with ID starting with eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984147 4959 scope.go:117] "RemoveContainer" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984447 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} err="failed to get container status \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": rpc error: code = NotFound desc = could not find container \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": container with ID starting with c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984470 4959 scope.go:117] "RemoveContainer" containerID="1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984640 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546"} err="failed to get container status \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": rpc error: code = NotFound desc = could not find container \"1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546\": container with ID starting with 1e87e5017bf5896baa8c9e3a92cef03952a1005dc8bca0720a87a5deec5bc546 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984663 4959 scope.go:117] "RemoveContainer" containerID="35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984954 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1"} err="failed to get container status \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": rpc error: code = NotFound desc = could not find container \"35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1\": container with ID starting with 35b09c8b30f9493f9fee1c5592be36e6780e27cf2358bbfa0f78c20aa21562b1 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.984979 4959 scope.go:117] "RemoveContainer" containerID="2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985181 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b"} err="failed to get container status \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": rpc error: code = NotFound desc = could not find container \"2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b\": container with ID starting with 2473e5554aebb4860b5f7b37bcc90606c5a8ca79e33f12fe5a839894e038b48b not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985249 4959 scope.go:117] "RemoveContainer" containerID="023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985414 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab"} err="failed to get container status \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": rpc error: code = NotFound desc = could not find container \"023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab\": container with ID starting with 023e35dcab822fd0fc85a179697db625c4ab0406be8421e5f28676956fdc8fab not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985433 4959 scope.go:117] "RemoveContainer" containerID="303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985712 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15"} err="failed to get container status \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": rpc error: code = NotFound desc = could not find container \"303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15\": container with ID starting with 303ccf287672f74e671764148ca9a095f3b1f68c588fc08801217f479896aa15 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985730 4959 scope.go:117] "RemoveContainer" containerID="c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985889 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2"} err="failed to get container status \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": rpc error: code = NotFound desc = could not find container \"c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2\": container with ID starting with c42a2026d32310d8ffd9a632af156d62469eee558064b11cc58383a06bfb2da2 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.985911 4959 scope.go:117] "RemoveContainer" containerID="cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986166 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd"} err="failed to get container status \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": rpc error: code = NotFound desc = could not find container \"cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd\": container with ID starting with cdff3006d6ce005d98c5aae8bcd7a80d75ea4d37e5e41d0e4b6ef65f663a45fd not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986229 4959 scope.go:117] "RemoveContainer" containerID="c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986451 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8"} err="failed to get container status \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": rpc error: code = NotFound desc = could not find container \"c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8\": container with ID starting with c0afab9e55dda864002c269f2befb4512b626a439fe982f5b5305e5a6c349ea8 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986471 4959 scope.go:117] "RemoveContainer" containerID="eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986657 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47"} err="failed to get container status \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": rpc error: code = NotFound desc = could not find container \"eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47\": container with ID starting with eb0244d3cc1684e32d1466ae57fe017f0a5626c1cabcc545fd9c16a5ae354c47 not found: ID does not exist" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.986689 4959 scope.go:117] "RemoveContainer" containerID="c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598" Oct 03 13:44:45 crc kubenswrapper[4959]: I1003 13:44:45.987056 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598"} err="failed to get container status \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": rpc error: code = NotFound desc = could not find container \"c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598\": container with ID starting with c66d1c9d924b8b0d328bf5b4a3c5039f6340ec807b7a661178c584da4c87f598 not found: ID does not exist" Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.693378 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tfz2x_117b68eb-14c1-456d-8c37-6259f7328587/kube-multus/2.log" Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.693690 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tfz2x" event={"ID":"117b68eb-14c1-456d-8c37-6259f7328587","Type":"ContainerStarted","Data":"58b3a7e7a4d27ecb5ca130deb2023eb6a9da74f42867e14f99c8f0d43c195952"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"66420e015830d8c93e189012138f1af1daf452c52ae0ae683e1ce38a0c4a4376"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698641 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"d8672e7caf78f327399ebca638786925d24b6a9f717d0c16e4f0556b3a0212a4"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698652 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"a28aec34f31fba6c29160de3fc6ad6476305c8c9feff09a8b9a0534b4ac585d9"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698661 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"62367bb6139d1a8b408f03cadcdb8dcdec4132ba1102153d67fdbf4fd5d3ecfe"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"c5564b33053fa48a2a0facab005c4d65c8de8ae15d1a767c66527a5ea579f0df"} Oct 03 13:44:46 crc kubenswrapper[4959]: I1003 13:44:46.698683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"b26d9b4089b5ef5348275f6a869e48d85730ab6b915560cfc7d8495dc762fe32"} Oct 03 13:44:47 crc kubenswrapper[4959]: I1003 13:44:47.699398 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3185a69a-b160-4257-afe1-a2165aa6e961" path="/var/lib/kubelet/pods/3185a69a-b160-4257-afe1-a2165aa6e961/volumes" Oct 03 13:44:49 crc kubenswrapper[4959]: I1003 13:44:49.719452 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"4d26fdfd75c953330136015dcd942bbd5f18801c5dfba1dc604a1fa73a663600"} Oct 03 13:44:51 crc kubenswrapper[4959]: I1003 13:44:51.731938 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" event={"ID":"401139a0-a614-4077-bbfd-e62293a96516","Type":"ContainerStarted","Data":"e4d806554b9fecfd88916cc93030ae4eeec22a68e1e0d8f4559284d6994bda13"} Oct 03 13:44:51 crc kubenswrapper[4959]: I1003 13:44:51.733496 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:51 crc kubenswrapper[4959]: I1003 13:44:51.761257 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" podStartSLOduration=6.76123646 podStartE2EDuration="6.76123646s" podCreationTimestamp="2025-10-03 13:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:44:51.757173774 +0000 UTC m=+860.960517211" watchObservedRunningTime="2025-10-03 13:44:51.76123646 +0000 UTC m=+860.964579887" Oct 03 13:44:51 crc kubenswrapper[4959]: I1003 13:44:51.763607 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:52 crc kubenswrapper[4959]: I1003 13:44:52.740799 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:52 crc kubenswrapper[4959]: I1003 13:44:52.740844 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:52 crc kubenswrapper[4959]: I1003 13:44:52.774478 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.251137 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-h5xbj"] Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.251812 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.253538 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.253575 4959 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zbx7v" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.253971 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.254522 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.254825 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-h5xbj"] Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.347323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2sz9\" (UniqueName: \"kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.347407 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.347459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.448595 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.448873 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2sz9\" (UniqueName: \"kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.449001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.449222 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.449393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.466905 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2sz9\" (UniqueName: \"kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9\") pod \"crc-storage-crc-h5xbj\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.583617 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.625055 4959 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(e396dc9689383ccfac756f47bf7db65d89e3165fe051f52958af5b8b78610411): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.625136 4959 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(e396dc9689383ccfac756f47bf7db65d89e3165fe051f52958af5b8b78610411): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.625163 4959 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(e396dc9689383ccfac756f47bf7db65d89e3165fe051f52958af5b8b78610411): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.625278 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-h5xbj_crc-storage(c19b5b78-580c-45b0-84a7-5fb0460a0114)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-h5xbj_crc-storage(c19b5b78-580c-45b0-84a7-5fb0460a0114)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(e396dc9689383ccfac756f47bf7db65d89e3165fe051f52958af5b8b78610411): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-h5xbj" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.744986 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: I1003 13:44:53.745894 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.767469 4959 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(654fc81d2892609d828a36a4874c14033f0297ceb46efda739b78d9efe87d175): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.767541 4959 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(654fc81d2892609d828a36a4874c14033f0297ceb46efda739b78d9efe87d175): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.767566 4959 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(654fc81d2892609d828a36a4874c14033f0297ceb46efda739b78d9efe87d175): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:44:53 crc kubenswrapper[4959]: E1003 13:44:53.767610 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-h5xbj_crc-storage(c19b5b78-580c-45b0-84a7-5fb0460a0114)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-h5xbj_crc-storage(c19b5b78-580c-45b0-84a7-5fb0460a0114)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-h5xbj_crc-storage_c19b5b78-580c-45b0-84a7-5fb0460a0114_0(654fc81d2892609d828a36a4874c14033f0297ceb46efda739b78d9efe87d175): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-h5xbj" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.130853 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm"] Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.131889 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.142426 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.142935 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.150074 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm"] Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.331000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.331048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.331074 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctnwh\" (UniqueName: \"kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.431888 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.432695 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.432789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctnwh\" (UniqueName: \"kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.433561 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.440232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.463882 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctnwh\" (UniqueName: \"kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh\") pod \"collect-profiles-29324985-mdznm\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.752136 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:00 crc kubenswrapper[4959]: I1003 13:45:00.925114 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm"] Oct 03 13:45:01 crc kubenswrapper[4959]: I1003 13:45:01.787829 4959 generic.go:334] "Generic (PLEG): container finished" podID="1e9c3104-76bd-4252-9901-37e9d6d14632" containerID="2aeef83766ab49120fc33f5fe8c4e31f870ab2e9e8bcffe8be170d4434c8caff" exitCode=0 Oct 03 13:45:01 crc kubenswrapper[4959]: I1003 13:45:01.787934 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" event={"ID":"1e9c3104-76bd-4252-9901-37e9d6d14632","Type":"ContainerDied","Data":"2aeef83766ab49120fc33f5fe8c4e31f870ab2e9e8bcffe8be170d4434c8caff"} Oct 03 13:45:01 crc kubenswrapper[4959]: I1003 13:45:01.788129 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" event={"ID":"1e9c3104-76bd-4252-9901-37e9d6d14632","Type":"ContainerStarted","Data":"e1fdb19f4aa67f97dc38410c1276c8c207d4bd0e9f351fde5ba7041ad9a13ff0"} Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.016070 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.061786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctnwh\" (UniqueName: \"kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh\") pod \"1e9c3104-76bd-4252-9901-37e9d6d14632\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.061859 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume\") pod \"1e9c3104-76bd-4252-9901-37e9d6d14632\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.061885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume\") pod \"1e9c3104-76bd-4252-9901-37e9d6d14632\" (UID: \"1e9c3104-76bd-4252-9901-37e9d6d14632\") " Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.062712 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e9c3104-76bd-4252-9901-37e9d6d14632" (UID: "1e9c3104-76bd-4252-9901-37e9d6d14632"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.067612 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e9c3104-76bd-4252-9901-37e9d6d14632" (UID: "1e9c3104-76bd-4252-9901-37e9d6d14632"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.068045 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh" (OuterVolumeSpecName: "kube-api-access-ctnwh") pod "1e9c3104-76bd-4252-9901-37e9d6d14632" (UID: "1e9c3104-76bd-4252-9901-37e9d6d14632"). InnerVolumeSpecName "kube-api-access-ctnwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.162949 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctnwh\" (UniqueName: \"kubernetes.io/projected/1e9c3104-76bd-4252-9901-37e9d6d14632-kube-api-access-ctnwh\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.162978 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e9c3104-76bd-4252-9901-37e9d6d14632-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.162987 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9c3104-76bd-4252-9901-37e9d6d14632-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.799855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" event={"ID":"1e9c3104-76bd-4252-9901-37e9d6d14632","Type":"ContainerDied","Data":"e1fdb19f4aa67f97dc38410c1276c8c207d4bd0e9f351fde5ba7041ad9a13ff0"} Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.799899 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1fdb19f4aa67f97dc38410c1276c8c207d4bd0e9f351fde5ba7041ad9a13ff0" Oct 03 13:45:03 crc kubenswrapper[4959]: I1003 13:45:03.799920 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm" Oct 03 13:45:07 crc kubenswrapper[4959]: I1003 13:45:07.685178 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:45:07 crc kubenswrapper[4959]: I1003 13:45:07.685855 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:45:07 crc kubenswrapper[4959]: I1003 13:45:07.883452 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-h5xbj"] Oct 03 13:45:08 crc kubenswrapper[4959]: I1003 13:45:08.837527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h5xbj" event={"ID":"c19b5b78-580c-45b0-84a7-5fb0460a0114","Type":"ContainerStarted","Data":"cbee049848ed4eb9c2d128345d89f6c59a42df68f6be78fdb69b2b385a6ad651"} Oct 03 13:45:09 crc kubenswrapper[4959]: I1003 13:45:09.847989 4959 generic.go:334] "Generic (PLEG): container finished" podID="c19b5b78-580c-45b0-84a7-5fb0460a0114" containerID="2f78b669c90a683d26ba998dad174b1a18caeba81dcd7bda93930601ce904512" exitCode=0 Oct 03 13:45:09 crc kubenswrapper[4959]: I1003 13:45:09.848100 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h5xbj" event={"ID":"c19b5b78-580c-45b0-84a7-5fb0460a0114","Type":"ContainerDied","Data":"2f78b669c90a683d26ba998dad174b1a18caeba81dcd7bda93930601ce904512"} Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.060371 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.262932 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage\") pod \"c19b5b78-580c-45b0-84a7-5fb0460a0114\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.263025 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2sz9\" (UniqueName: \"kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9\") pod \"c19b5b78-580c-45b0-84a7-5fb0460a0114\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.263086 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt\") pod \"c19b5b78-580c-45b0-84a7-5fb0460a0114\" (UID: \"c19b5b78-580c-45b0-84a7-5fb0460a0114\") " Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.263333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "c19b5b78-580c-45b0-84a7-5fb0460a0114" (UID: "c19b5b78-580c-45b0-84a7-5fb0460a0114"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.263512 4959 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/c19b5b78-580c-45b0-84a7-5fb0460a0114-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.271186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9" (OuterVolumeSpecName: "kube-api-access-s2sz9") pod "c19b5b78-580c-45b0-84a7-5fb0460a0114" (UID: "c19b5b78-580c-45b0-84a7-5fb0460a0114"). InnerVolumeSpecName "kube-api-access-s2sz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.281351 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "c19b5b78-580c-45b0-84a7-5fb0460a0114" (UID: "c19b5b78-580c-45b0-84a7-5fb0460a0114"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.364325 4959 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/c19b5b78-580c-45b0-84a7-5fb0460a0114-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.364386 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2sz9\" (UniqueName: \"kubernetes.io/projected/c19b5b78-580c-45b0-84a7-5fb0460a0114-kube-api-access-s2sz9\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.859705 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h5xbj" event={"ID":"c19b5b78-580c-45b0-84a7-5fb0460a0114","Type":"ContainerDied","Data":"cbee049848ed4eb9c2d128345d89f6c59a42df68f6be78fdb69b2b385a6ad651"} Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.859759 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbee049848ed4eb9c2d128345d89f6c59a42df68f6be78fdb69b2b385a6ad651" Oct 03 13:45:11 crc kubenswrapper[4959]: I1003 13:45:11.859782 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h5xbj" Oct 03 13:45:15 crc kubenswrapper[4959]: I1003 13:45:15.375436 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-srzq5" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.877492 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7"] Oct 03 13:45:17 crc kubenswrapper[4959]: E1003 13:45:17.877958 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e9c3104-76bd-4252-9901-37e9d6d14632" containerName="collect-profiles" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.877973 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e9c3104-76bd-4252-9901-37e9d6d14632" containerName="collect-profiles" Oct 03 13:45:17 crc kubenswrapper[4959]: E1003 13:45:17.877989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" containerName="storage" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.877995 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" containerName="storage" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.878097 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e9c3104-76bd-4252-9901-37e9d6d14632" containerName="collect-profiles" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.878108 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" containerName="storage" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.878878 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.880629 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.885894 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7"] Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.947734 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjt72\" (UniqueName: \"kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.947771 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:17 crc kubenswrapper[4959]: I1003 13:45:17.947807 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.048904 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjt72\" (UniqueName: \"kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.049241 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.049343 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.049752 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.049852 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.071341 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjt72\" (UniqueName: \"kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.200071 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.385416 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7"] Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.897265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerStarted","Data":"6bd77ec75ec22d0c49f7d1f47e14964d32e2710f602a9b6ca0dead48f3449cb6"} Oct 03 13:45:18 crc kubenswrapper[4959]: I1003 13:45:18.897583 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerStarted","Data":"95c30d79459a15333f0a20c24f4b9abdc0b047d32a1d6890c84a41ae98b57f16"} Oct 03 13:45:19 crc kubenswrapper[4959]: I1003 13:45:19.902860 4959 generic.go:334] "Generic (PLEG): container finished" podID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerID="6bd77ec75ec22d0c49f7d1f47e14964d32e2710f602a9b6ca0dead48f3449cb6" exitCode=0 Oct 03 13:45:19 crc kubenswrapper[4959]: I1003 13:45:19.902905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerDied","Data":"6bd77ec75ec22d0c49f7d1f47e14964d32e2710f602a9b6ca0dead48f3449cb6"} Oct 03 13:45:21 crc kubenswrapper[4959]: I1003 13:45:21.917340 4959 generic.go:334] "Generic (PLEG): container finished" podID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerID="786f4362af36c0bc117bf4c63d3f05bc09d1101e0caba6b2397d8c7ed2463bd7" exitCode=0 Oct 03 13:45:21 crc kubenswrapper[4959]: I1003 13:45:21.917473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerDied","Data":"786f4362af36c0bc117bf4c63d3f05bc09d1101e0caba6b2397d8c7ed2463bd7"} Oct 03 13:45:22 crc kubenswrapper[4959]: I1003 13:45:22.927056 4959 generic.go:334] "Generic (PLEG): container finished" podID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerID="1ab6de6a7da54232a0f9abbbc330d62e407dbeb05e05d64aac9d81b6809d4de8" exitCode=0 Oct 03 13:45:22 crc kubenswrapper[4959]: I1003 13:45:22.927133 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerDied","Data":"1ab6de6a7da54232a0f9abbbc330d62e407dbeb05e05d64aac9d81b6809d4de8"} Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.201408 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.226391 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjt72\" (UniqueName: \"kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72\") pod \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.226494 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util\") pod \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.226583 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle\") pod \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\" (UID: \"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0\") " Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.227871 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle" (OuterVolumeSpecName: "bundle") pod "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" (UID: "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.243278 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72" (OuterVolumeSpecName: "kube-api-access-vjt72") pod "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" (UID: "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0"). InnerVolumeSpecName "kube-api-access-vjt72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.243531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util" (OuterVolumeSpecName: "util") pod "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" (UID: "ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.328052 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.328095 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjt72\" (UniqueName: \"kubernetes.io/projected/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-kube-api-access-vjt72\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.328104 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.941781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" event={"ID":"ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0","Type":"ContainerDied","Data":"95c30d79459a15333f0a20c24f4b9abdc0b047d32a1d6890c84a41ae98b57f16"} Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.941844 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95c30d79459a15333f0a20c24f4b9abdc0b047d32a1d6890c84a41ae98b57f16" Oct 03 13:45:24 crc kubenswrapper[4959]: I1003 13:45:24.941842 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.676182 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q97bx"] Oct 03 13:45:28 crc kubenswrapper[4959]: E1003 13:45:28.677349 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="extract" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.677416 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="extract" Oct 03 13:45:28 crc kubenswrapper[4959]: E1003 13:45:28.677513 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="util" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.677564 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="util" Oct 03 13:45:28 crc kubenswrapper[4959]: E1003 13:45:28.677615 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="pull" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.677662 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="pull" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.677805 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0" containerName="extract" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.678182 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.678675 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xrdp\" (UniqueName: \"kubernetes.io/projected/92c271a8-8637-4497-aa97-f63e1d9134b4-kube-api-access-7xrdp\") pod \"nmstate-operator-858ddd8f98-q97bx\" (UID: \"92c271a8-8637-4497-aa97-f63e1d9134b4\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.686718 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.686876 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.687342 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-tm8bs" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.700592 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q97bx"] Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.780016 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xrdp\" (UniqueName: \"kubernetes.io/projected/92c271a8-8637-4497-aa97-f63e1d9134b4-kube-api-access-7xrdp\") pod \"nmstate-operator-858ddd8f98-q97bx\" (UID: \"92c271a8-8637-4497-aa97-f63e1d9134b4\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.797693 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xrdp\" (UniqueName: \"kubernetes.io/projected/92c271a8-8637-4497-aa97-f63e1d9134b4-kube-api-access-7xrdp\") pod \"nmstate-operator-858ddd8f98-q97bx\" (UID: \"92c271a8-8637-4497-aa97-f63e1d9134b4\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" Oct 03 13:45:28 crc kubenswrapper[4959]: I1003 13:45:28.994951 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" Oct 03 13:45:29 crc kubenswrapper[4959]: I1003 13:45:29.191719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q97bx"] Oct 03 13:45:29 crc kubenswrapper[4959]: W1003 13:45:29.204393 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92c271a8_8637_4497_aa97_f63e1d9134b4.slice/crio-a4d7b40d969a8103eb534823734116a28dba20609115132a269a05d97f821387 WatchSource:0}: Error finding container a4d7b40d969a8103eb534823734116a28dba20609115132a269a05d97f821387: Status 404 returned error can't find the container with id a4d7b40d969a8103eb534823734116a28dba20609115132a269a05d97f821387 Oct 03 13:45:29 crc kubenswrapper[4959]: I1003 13:45:29.967971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" event={"ID":"92c271a8-8637-4497-aa97-f63e1d9134b4","Type":"ContainerStarted","Data":"a4d7b40d969a8103eb534823734116a28dba20609115132a269a05d97f821387"} Oct 03 13:45:31 crc kubenswrapper[4959]: I1003 13:45:31.977094 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" event={"ID":"92c271a8-8637-4497-aa97-f63e1d9134b4","Type":"ContainerStarted","Data":"0bc325a655f02ba9b5fd0b612c2297b58cd7a0c9f5c47ee5129f9c17e66a3c4d"} Oct 03 13:45:32 crc kubenswrapper[4959]: I1003 13:45:32.002876 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q97bx" podStartSLOduration=2.046339834 podStartE2EDuration="4.002847519s" podCreationTimestamp="2025-10-03 13:45:28 +0000 UTC" firstStartedPulling="2025-10-03 13:45:29.206812373 +0000 UTC m=+898.410155780" lastFinishedPulling="2025-10-03 13:45:31.163320048 +0000 UTC m=+900.366663465" observedRunningTime="2025-10-03 13:45:31.996723145 +0000 UTC m=+901.200066592" watchObservedRunningTime="2025-10-03 13:45:32.002847519 +0000 UTC m=+901.206190976" Oct 03 13:45:36 crc kubenswrapper[4959]: I1003 13:45:36.044947 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:45:36 crc kubenswrapper[4959]: I1003 13:45:36.045012 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.389898 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.391133 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.392641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mbv2\" (UniqueName: \"kubernetes.io/projected/e70ee995-3820-471a-a3d9-c01e0eaa2ab1-kube-api-access-5mbv2\") pod \"nmstate-metrics-fdff9cb8d-zwrzs\" (UID: \"e70ee995-3820-471a-a3d9-c01e0eaa2ab1\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.393944 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-gdbvn" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.409706 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.411762 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.412594 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.422062 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.440831 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-j82kp"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.441701 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.446085 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493656 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-dbus-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493694 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-ovs-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493722 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hsnm\" (UniqueName: \"kubernetes.io/projected/954033ef-4a14-4151-b475-d7e97ad2a7c9-kube-api-access-4hsnm\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493763 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mbv2\" (UniqueName: \"kubernetes.io/projected/e70ee995-3820-471a-a3d9-c01e0eaa2ab1-kube-api-access-5mbv2\") pod \"nmstate-metrics-fdff9cb8d-zwrzs\" (UID: \"e70ee995-3820-471a-a3d9-c01e0eaa2ab1\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493794 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-nmstate-lock\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.493816 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-885gr\" (UniqueName: \"kubernetes.io/projected/48c1908a-76ca-4b8f-ab18-6993d6b402db-kube-api-access-885gr\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.515030 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mbv2\" (UniqueName: \"kubernetes.io/projected/e70ee995-3820-471a-a3d9-c01e0eaa2ab1-kube-api-access-5mbv2\") pod \"nmstate-metrics-fdff9cb8d-zwrzs\" (UID: \"e70ee995-3820-471a-a3d9-c01e0eaa2ab1\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.528292 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.533546 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.535972 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.536529 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.537340 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xl9np" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.537799 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-dbus-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594484 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-ovs-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594512 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594532 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hsnm\" (UniqueName: \"kubernetes.io/projected/954033ef-4a14-4151-b475-d7e97ad2a7c9-kube-api-access-4hsnm\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594559 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsfxz\" (UniqueName: \"kubernetes.io/projected/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-kube-api-access-gsfxz\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594576 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594585 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-ovs-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594595 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-nmstate-lock\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594614 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-nmstate-lock\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: E1003 13:45:38.594684 4959 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 03 13:45:38 crc kubenswrapper[4959]: E1003 13:45:38.594729 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair podName:48c1908a-76ca-4b8f-ab18-6993d6b402db nodeName:}" failed. No retries permitted until 2025-10-03 13:45:39.094714077 +0000 UTC m=+908.298057484 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair") pod "nmstate-webhook-6cdbc54649-mkvvt" (UID: "48c1908a-76ca-4b8f-ab18-6993d6b402db") : secret "openshift-nmstate-webhook" not found Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594790 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/954033ef-4a14-4151-b475-d7e97ad2a7c9-dbus-socket\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-885gr\" (UniqueName: \"kubernetes.io/projected/48c1908a-76ca-4b8f-ab18-6993d6b402db-kube-api-access-885gr\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.594885 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.614623 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hsnm\" (UniqueName: \"kubernetes.io/projected/954033ef-4a14-4151-b475-d7e97ad2a7c9-kube-api-access-4hsnm\") pod \"nmstate-handler-j82kp\" (UID: \"954033ef-4a14-4151-b475-d7e97ad2a7c9\") " pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.618098 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-885gr\" (UniqueName: \"kubernetes.io/projected/48c1908a-76ca-4b8f-ab18-6993d6b402db-kube-api-access-885gr\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.696269 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.696633 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.696707 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsfxz\" (UniqueName: \"kubernetes.io/projected/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-kube-api-access-gsfxz\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.698328 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.700792 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.719639 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-dfc8bb9b8-wpfbj"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.720331 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.722897 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.725997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsfxz\" (UniqueName: \"kubernetes.io/projected/4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a-kube-api-access-gsfxz\") pod \"nmstate-console-plugin-6b874cbd85-n5nlz\" (UID: \"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.740738 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dfc8bb9b8-wpfbj"] Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.769272 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.797759 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-oauth-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.797863 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-trusted-ca-bundle\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.797930 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrxll\" (UniqueName: \"kubernetes.io/projected/347055c0-aef7-42bf-8d91-f844932ea80a-kube-api-access-rrxll\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.797984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-service-ca\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.798031 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-oauth-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.798076 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.798092 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-console-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.855912 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-oauth-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900464 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-trusted-ca-bundle\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrxll\" (UniqueName: \"kubernetes.io/projected/347055c0-aef7-42bf-8d91-f844932ea80a-kube-api-access-rrxll\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900507 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-service-ca\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900528 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-oauth-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.900581 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-console-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.901238 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-oauth-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.902081 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-console-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.902140 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-service-ca\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.904163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/347055c0-aef7-42bf-8d91-f844932ea80a-trusted-ca-bundle\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.907830 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-serving-cert\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.910591 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/347055c0-aef7-42bf-8d91-f844932ea80a-console-oauth-config\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.923273 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrxll\" (UniqueName: \"kubernetes.io/projected/347055c0-aef7-42bf-8d91-f844932ea80a-kube-api-access-rrxll\") pod \"console-dfc8bb9b8-wpfbj\" (UID: \"347055c0-aef7-42bf-8d91-f844932ea80a\") " pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:38 crc kubenswrapper[4959]: I1003 13:45:38.926142 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs"] Oct 03 13:45:38 crc kubenswrapper[4959]: W1003 13:45:38.933229 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode70ee995_3820_471a_a3d9_c01e0eaa2ab1.slice/crio-c42cca675fe752a1e34471adde3d46a700e3c4b61ccbc0ebaa597ba87eaacb1f WatchSource:0}: Error finding container c42cca675fe752a1e34471adde3d46a700e3c4b61ccbc0ebaa597ba87eaacb1f: Status 404 returned error can't find the container with id c42cca675fe752a1e34471adde3d46a700e3c4b61ccbc0ebaa597ba87eaacb1f Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.010705 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" event={"ID":"e70ee995-3820-471a-a3d9-c01e0eaa2ab1","Type":"ContainerStarted","Data":"c42cca675fe752a1e34471adde3d46a700e3c4b61ccbc0ebaa597ba87eaacb1f"} Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.011502 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j82kp" event={"ID":"954033ef-4a14-4151-b475-d7e97ad2a7c9","Type":"ContainerStarted","Data":"2ad740210c46e87ea0ca934a4583cd0f67a4ba027f0894ad0e88b6f8e0faf89a"} Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.057801 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.102419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.107259 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/48c1908a-76ca-4b8f-ab18-6993d6b402db-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mkvvt\" (UID: \"48c1908a-76ca-4b8f-ab18-6993d6b402db\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.229666 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz"] Oct 03 13:45:39 crc kubenswrapper[4959]: W1003 13:45:39.235317 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dfc17f0_a6dc_4a02_91e3_4b9be95d6a4a.slice/crio-173246479d391b109d85a3a8c5e5d3baa0ddf494106580d084a7965833171af0 WatchSource:0}: Error finding container 173246479d391b109d85a3a8c5e5d3baa0ddf494106580d084a7965833171af0: Status 404 returned error can't find the container with id 173246479d391b109d85a3a8c5e5d3baa0ddf494106580d084a7965833171af0 Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.361217 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.455074 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-dfc8bb9b8-wpfbj"] Oct 03 13:45:39 crc kubenswrapper[4959]: W1003 13:45:39.463917 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod347055c0_aef7_42bf_8d91_f844932ea80a.slice/crio-3500437c05dfd5c67237b9aa0a4311c1610d938a4a1a3eff3e73184709bb2cc6 WatchSource:0}: Error finding container 3500437c05dfd5c67237b9aa0a4311c1610d938a4a1a3eff3e73184709bb2cc6: Status 404 returned error can't find the container with id 3500437c05dfd5c67237b9aa0a4311c1610d938a4a1a3eff3e73184709bb2cc6 Oct 03 13:45:39 crc kubenswrapper[4959]: I1003 13:45:39.747027 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt"] Oct 03 13:45:40 crc kubenswrapper[4959]: I1003 13:45:40.017642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" event={"ID":"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a","Type":"ContainerStarted","Data":"173246479d391b109d85a3a8c5e5d3baa0ddf494106580d084a7965833171af0"} Oct 03 13:45:40 crc kubenswrapper[4959]: I1003 13:45:40.018983 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dfc8bb9b8-wpfbj" event={"ID":"347055c0-aef7-42bf-8d91-f844932ea80a","Type":"ContainerStarted","Data":"97c88438e8d39679673f719e370f30f8bbb1b113532269b7451529d71b3c8d5c"} Oct 03 13:45:40 crc kubenswrapper[4959]: I1003 13:45:40.019025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-dfc8bb9b8-wpfbj" event={"ID":"347055c0-aef7-42bf-8d91-f844932ea80a","Type":"ContainerStarted","Data":"3500437c05dfd5c67237b9aa0a4311c1610d938a4a1a3eff3e73184709bb2cc6"} Oct 03 13:45:40 crc kubenswrapper[4959]: I1003 13:45:40.020848 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" event={"ID":"48c1908a-76ca-4b8f-ab18-6993d6b402db","Type":"ContainerStarted","Data":"e9a4bdd46efbc974e3b6a6bbde1d9413748fd9a509b4fbb3339c2a6031cd4962"} Oct 03 13:45:40 crc kubenswrapper[4959]: I1003 13:45:40.043281 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-dfc8bb9b8-wpfbj" podStartSLOduration=2.043259026 podStartE2EDuration="2.043259026s" podCreationTimestamp="2025-10-03 13:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:45:40.042604371 +0000 UTC m=+909.245947788" watchObservedRunningTime="2025-10-03 13:45:40.043259026 +0000 UTC m=+909.246602443" Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.032959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j82kp" event={"ID":"954033ef-4a14-4151-b475-d7e97ad2a7c9","Type":"ContainerStarted","Data":"539d94db341841f56a658ba35fd77658fd280487a160a018aac239bd7788085b"} Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.033517 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.034165 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" event={"ID":"e70ee995-3820-471a-a3d9-c01e0eaa2ab1","Type":"ContainerStarted","Data":"7b35e1bb16a5fc4f5c60e5c2ddcfc36eeb1938a52d8e0cebf553ad102bce97be"} Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.035677 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" event={"ID":"48c1908a-76ca-4b8f-ab18-6993d6b402db","Type":"ContainerStarted","Data":"95bb08ef41fd36e483a3d4cf0f0d362252c1e30feb4506405b0436bf012842e4"} Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.046825 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-j82kp" podStartSLOduration=1.688981585 podStartE2EDuration="4.046788855s" podCreationTimestamp="2025-10-03 13:45:38 +0000 UTC" firstStartedPulling="2025-10-03 13:45:38.788394949 +0000 UTC m=+907.991738366" lastFinishedPulling="2025-10-03 13:45:41.146202219 +0000 UTC m=+910.349545636" observedRunningTime="2025-10-03 13:45:42.045994115 +0000 UTC m=+911.249337532" watchObservedRunningTime="2025-10-03 13:45:42.046788855 +0000 UTC m=+911.250132272" Oct 03 13:45:42 crc kubenswrapper[4959]: I1003 13:45:42.067357 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" podStartSLOduration=2.700846113 podStartE2EDuration="4.067337341s" podCreationTimestamp="2025-10-03 13:45:38 +0000 UTC" firstStartedPulling="2025-10-03 13:45:39.753679035 +0000 UTC m=+908.957022452" lastFinishedPulling="2025-10-03 13:45:41.120170263 +0000 UTC m=+910.323513680" observedRunningTime="2025-10-03 13:45:42.063783617 +0000 UTC m=+911.267127044" watchObservedRunningTime="2025-10-03 13:45:42.067337341 +0000 UTC m=+911.270680758" Oct 03 13:45:43 crc kubenswrapper[4959]: I1003 13:45:43.042688 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" event={"ID":"4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a","Type":"ContainerStarted","Data":"64981e57524031253f02ce95d7c185d0db0ff2a606659e1b0603a86ec017aaea"} Oct 03 13:45:43 crc kubenswrapper[4959]: I1003 13:45:43.043040 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:45:43 crc kubenswrapper[4959]: I1003 13:45:43.060481 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-n5nlz" podStartSLOduration=2.210976894 podStartE2EDuration="5.060462636s" podCreationTimestamp="2025-10-03 13:45:38 +0000 UTC" firstStartedPulling="2025-10-03 13:45:39.238277502 +0000 UTC m=+908.441620919" lastFinishedPulling="2025-10-03 13:45:42.087763234 +0000 UTC m=+911.291106661" observedRunningTime="2025-10-03 13:45:43.059913603 +0000 UTC m=+912.263257020" watchObservedRunningTime="2025-10-03 13:45:43.060462636 +0000 UTC m=+912.263806043" Oct 03 13:45:44 crc kubenswrapper[4959]: I1003 13:45:44.050584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" event={"ID":"e70ee995-3820-471a-a3d9-c01e0eaa2ab1","Type":"ContainerStarted","Data":"f6ad107b07b5e131b4df21fcbe693958325372fed7414f6235190616b8c20581"} Oct 03 13:45:44 crc kubenswrapper[4959]: I1003 13:45:44.068832 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-zwrzs" podStartSLOduration=1.342311063 podStartE2EDuration="6.068809871s" podCreationTimestamp="2025-10-03 13:45:38 +0000 UTC" firstStartedPulling="2025-10-03 13:45:38.934954596 +0000 UTC m=+908.138298013" lastFinishedPulling="2025-10-03 13:45:43.661453404 +0000 UTC m=+912.864796821" observedRunningTime="2025-10-03 13:45:44.065969874 +0000 UTC m=+913.269313301" watchObservedRunningTime="2025-10-03 13:45:44.068809871 +0000 UTC m=+913.272153288" Oct 03 13:45:48 crc kubenswrapper[4959]: I1003 13:45:48.793275 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-j82kp" Oct 03 13:45:49 crc kubenswrapper[4959]: I1003 13:45:49.058559 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:49 crc kubenswrapper[4959]: I1003 13:45:49.058620 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:49 crc kubenswrapper[4959]: I1003 13:45:49.066573 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:49 crc kubenswrapper[4959]: I1003 13:45:49.084143 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-dfc8bb9b8-wpfbj" Oct 03 13:45:49 crc kubenswrapper[4959]: I1003 13:45:49.147847 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:45:59 crc kubenswrapper[4959]: I1003 13:45:59.368748 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mkvvt" Oct 03 13:46:06 crc kubenswrapper[4959]: I1003 13:46:06.044302 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:46:06 crc kubenswrapper[4959]: I1003 13:46:06.044771 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.642683 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5"] Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.645023 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.646898 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.651711 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5"] Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.801094 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.801131 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.801547 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pt9j\" (UniqueName: \"kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.902636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pt9j\" (UniqueName: \"kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.902809 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.902859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.903408 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.903583 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.927299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pt9j\" (UniqueName: \"kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:11 crc kubenswrapper[4959]: I1003 13:46:11.962902 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:12 crc kubenswrapper[4959]: I1003 13:46:12.356719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5"] Oct 03 13:46:13 crc kubenswrapper[4959]: I1003 13:46:13.208748 4959 generic.go:334] "Generic (PLEG): container finished" podID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerID="506cff558ecdff4215e41507748e67ae72a07404c3a815e47a01eccbad5fcf5e" exitCode=0 Oct 03 13:46:13 crc kubenswrapper[4959]: I1003 13:46:13.209101 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerDied","Data":"506cff558ecdff4215e41507748e67ae72a07404c3a815e47a01eccbad5fcf5e"} Oct 03 13:46:13 crc kubenswrapper[4959]: I1003 13:46:13.209733 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerStarted","Data":"75f5bb50a2fe524ce560f2df058ba247f26b0d6fd5f63cf03678cc475632c511"} Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.203969 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-l2fnx" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" containerID="cri-o://e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8" gracePeriod=15 Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.593230 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l2fnx_819c5b6d-500f-4954-8225-a229c415bca9/console/0.log" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.593308 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638394 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638485 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk6mc\" (UniqueName: \"kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638502 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638541 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638581 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638631 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.638653 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config\") pod \"819c5b6d-500f-4954-8225-a229c415bca9\" (UID: \"819c5b6d-500f-4954-8225-a229c415bca9\") " Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.639353 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca" (OuterVolumeSpecName: "service-ca") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.639394 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config" (OuterVolumeSpecName: "console-config") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.639405 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.639361 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.643652 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.643668 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc" (OuterVolumeSpecName: "kube-api-access-bk6mc") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "kube-api-access-bk6mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.643922 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "819c5b6d-500f-4954-8225-a229c415bca9" (UID: "819c5b6d-500f-4954-8225-a229c415bca9"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740054 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk6mc\" (UniqueName: \"kubernetes.io/projected/819c5b6d-500f-4954-8225-a229c415bca9-kube-api-access-bk6mc\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740095 4959 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740104 4959 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740112 4959 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740121 4959 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740129 4959 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/819c5b6d-500f-4954-8225-a229c415bca9-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:14 crc kubenswrapper[4959]: I1003 13:46:14.740137 4959 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/819c5b6d-500f-4954-8225-a229c415bca9-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.222280 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l2fnx_819c5b6d-500f-4954-8225-a229c415bca9/console/0.log" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.222774 4959 generic.go:334] "Generic (PLEG): container finished" podID="819c5b6d-500f-4954-8225-a229c415bca9" containerID="e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8" exitCode=2 Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.222867 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l2fnx" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.222858 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l2fnx" event={"ID":"819c5b6d-500f-4954-8225-a229c415bca9","Type":"ContainerDied","Data":"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8"} Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.223067 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l2fnx" event={"ID":"819c5b6d-500f-4954-8225-a229c415bca9","Type":"ContainerDied","Data":"2d694048f15b02f41a4e83e23e00c360fb33a6e412dcfe58259275a95189a09e"} Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.223172 4959 scope.go:117] "RemoveContainer" containerID="e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.224853 4959 generic.go:334] "Generic (PLEG): container finished" podID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerID="0ac3c84e34c04574c7fad9e13d6520050eb1bf9250e76c01cd55e26c8e0b4038" exitCode=0 Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.224934 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerDied","Data":"0ac3c84e34c04574c7fad9e13d6520050eb1bf9250e76c01cd55e26c8e0b4038"} Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.257600 4959 scope.go:117] "RemoveContainer" containerID="e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8" Oct 03 13:46:15 crc kubenswrapper[4959]: E1003 13:46:15.258695 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8\": container with ID starting with e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8 not found: ID does not exist" containerID="e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.258733 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8"} err="failed to get container status \"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8\": rpc error: code = NotFound desc = could not find container \"e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8\": container with ID starting with e18e0a39cfdc0595050486e22bc78a42c3878c62dfdcb980b703e0a90518cda8 not found: ID does not exist" Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.260084 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.264962 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-l2fnx"] Oct 03 13:46:15 crc kubenswrapper[4959]: I1003 13:46:15.694623 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819c5b6d-500f-4954-8225-a229c415bca9" path="/var/lib/kubelet/pods/819c5b6d-500f-4954-8225-a229c415bca9/volumes" Oct 03 13:46:16 crc kubenswrapper[4959]: I1003 13:46:16.234314 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerStarted","Data":"a8d099c16f81a90c23b31f69ab3a9c92aa2c403ec47665b2ba76ec4c7c3d3766"} Oct 03 13:46:16 crc kubenswrapper[4959]: I1003 13:46:16.252429 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" podStartSLOduration=3.6336922339999997 podStartE2EDuration="5.252412919s" podCreationTimestamp="2025-10-03 13:46:11 +0000 UTC" firstStartedPulling="2025-10-03 13:46:13.210258418 +0000 UTC m=+942.413601835" lastFinishedPulling="2025-10-03 13:46:14.828979113 +0000 UTC m=+944.032322520" observedRunningTime="2025-10-03 13:46:16.25032507 +0000 UTC m=+945.453668497" watchObservedRunningTime="2025-10-03 13:46:16.252412919 +0000 UTC m=+945.455756336" Oct 03 13:46:17 crc kubenswrapper[4959]: I1003 13:46:17.242127 4959 generic.go:334] "Generic (PLEG): container finished" podID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerID="a8d099c16f81a90c23b31f69ab3a9c92aa2c403ec47665b2ba76ec4c7c3d3766" exitCode=0 Oct 03 13:46:17 crc kubenswrapper[4959]: I1003 13:46:17.242239 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerDied","Data":"a8d099c16f81a90c23b31f69ab3a9c92aa2c403ec47665b2ba76ec4c7c3d3766"} Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.454131 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.493722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle\") pod \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.493797 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pt9j\" (UniqueName: \"kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j\") pod \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.493869 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util\") pod \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\" (UID: \"b76d5d80-bfab-404a-9b13-ab54b5bf0a30\") " Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.494875 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle" (OuterVolumeSpecName: "bundle") pod "b76d5d80-bfab-404a-9b13-ab54b5bf0a30" (UID: "b76d5d80-bfab-404a-9b13-ab54b5bf0a30"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.499505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j" (OuterVolumeSpecName: "kube-api-access-6pt9j") pod "b76d5d80-bfab-404a-9b13-ab54b5bf0a30" (UID: "b76d5d80-bfab-404a-9b13-ab54b5bf0a30"). InnerVolumeSpecName "kube-api-access-6pt9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.503608 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util" (OuterVolumeSpecName: "util") pod "b76d5d80-bfab-404a-9b13-ab54b5bf0a30" (UID: "b76d5d80-bfab-404a-9b13-ab54b5bf0a30"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.595495 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.595535 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pt9j\" (UniqueName: \"kubernetes.io/projected/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-kube-api-access-6pt9j\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:18 crc kubenswrapper[4959]: I1003 13:46:18.595551 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b76d5d80-bfab-404a-9b13-ab54b5bf0a30-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:46:19 crc kubenswrapper[4959]: I1003 13:46:19.255940 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" event={"ID":"b76d5d80-bfab-404a-9b13-ab54b5bf0a30","Type":"ContainerDied","Data":"75f5bb50a2fe524ce560f2df058ba247f26b0d6fd5f63cf03678cc475632c511"} Oct 03 13:46:19 crc kubenswrapper[4959]: I1003 13:46:19.256006 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f5bb50a2fe524ce560f2df058ba247f26b0d6fd5f63cf03678cc475632c511" Oct 03 13:46:19 crc kubenswrapper[4959]: I1003 13:46:19.256014 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.233861 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx"] Oct 03 13:46:28 crc kubenswrapper[4959]: E1003 13:46:28.234572 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="extract" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234587 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="extract" Oct 03 13:46:28 crc kubenswrapper[4959]: E1003 13:46:28.234608 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234614 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" Oct 03 13:46:28 crc kubenswrapper[4959]: E1003 13:46:28.234627 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="pull" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234633 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="pull" Oct 03 13:46:28 crc kubenswrapper[4959]: E1003 13:46:28.234642 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="util" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234647 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="util" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234737 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="819c5b6d-500f-4954-8225-a229c415bca9" containerName="console" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.234748 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b76d5d80-bfab-404a-9b13-ab54b5bf0a30" containerName="extract" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.235086 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.236833 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.237273 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.237312 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gcxsc" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.237918 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.240524 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.257948 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx"] Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.311991 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-webhook-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.312047 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4twkw\" (UniqueName: \"kubernetes.io/projected/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-kube-api-access-4twkw\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.312167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-apiservice-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.413054 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4twkw\" (UniqueName: \"kubernetes.io/projected/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-kube-api-access-4twkw\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.413148 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-apiservice-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.413233 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-webhook-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.427705 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-webhook-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.427712 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-apiservice-cert\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.448465 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4twkw\" (UniqueName: \"kubernetes.io/projected/7e4e41b5-dc12-4552-a5f2-28b01e76b84d-kube-api-access-4twkw\") pod \"metallb-operator-controller-manager-fd787bc66-gxclx\" (UID: \"7e4e41b5-dc12-4552-a5f2-28b01e76b84d\") " pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.552327 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.670322 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp"] Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.678646 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.680112 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp"] Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.681371 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-wxcmf" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.681564 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.681681 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.817898 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-webhook-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.817948 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-apiservice-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.817983 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dddqs\" (UniqueName: \"kubernetes.io/projected/32ace841-e798-4101-8a6a-a7f574ee693a-kube-api-access-dddqs\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.919569 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-webhook-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.919620 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-apiservice-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.919660 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dddqs\" (UniqueName: \"kubernetes.io/projected/32ace841-e798-4101-8a6a-a7f574ee693a-kube-api-access-dddqs\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.924817 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-apiservice-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.924875 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32ace841-e798-4101-8a6a-a7f574ee693a-webhook-cert\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:28 crc kubenswrapper[4959]: I1003 13:46:28.938162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dddqs\" (UniqueName: \"kubernetes.io/projected/32ace841-e798-4101-8a6a-a7f574ee693a-kube-api-access-dddqs\") pod \"metallb-operator-webhook-server-98cf898b6-fmszp\" (UID: \"32ace841-e798-4101-8a6a-a7f574ee693a\") " pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:29 crc kubenswrapper[4959]: I1003 13:46:29.020233 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:29 crc kubenswrapper[4959]: I1003 13:46:29.094616 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx"] Oct 03 13:46:29 crc kubenswrapper[4959]: W1003 13:46:29.106280 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e4e41b5_dc12_4552_a5f2_28b01e76b84d.slice/crio-877ba4189eaef52215eaa45e0037f3f5caebc3638030899fe337e74b36eaafc7 WatchSource:0}: Error finding container 877ba4189eaef52215eaa45e0037f3f5caebc3638030899fe337e74b36eaafc7: Status 404 returned error can't find the container with id 877ba4189eaef52215eaa45e0037f3f5caebc3638030899fe337e74b36eaafc7 Oct 03 13:46:29 crc kubenswrapper[4959]: I1003 13:46:29.242937 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp"] Oct 03 13:46:29 crc kubenswrapper[4959]: W1003 13:46:29.247924 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ace841_e798_4101_8a6a_a7f574ee693a.slice/crio-e7716d9530a5ebbedf9c7a015fa885b583c32fd8d2db3934aeea779d3911c434 WatchSource:0}: Error finding container e7716d9530a5ebbedf9c7a015fa885b583c32fd8d2db3934aeea779d3911c434: Status 404 returned error can't find the container with id e7716d9530a5ebbedf9c7a015fa885b583c32fd8d2db3934aeea779d3911c434 Oct 03 13:46:29 crc kubenswrapper[4959]: I1003 13:46:29.310590 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" event={"ID":"7e4e41b5-dc12-4552-a5f2-28b01e76b84d","Type":"ContainerStarted","Data":"877ba4189eaef52215eaa45e0037f3f5caebc3638030899fe337e74b36eaafc7"} Oct 03 13:46:29 crc kubenswrapper[4959]: I1003 13:46:29.311546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" event={"ID":"32ace841-e798-4101-8a6a-a7f574ee693a","Type":"ContainerStarted","Data":"e7716d9530a5ebbedf9c7a015fa885b583c32fd8d2db3934aeea779d3911c434"} Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.341778 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" event={"ID":"32ace841-e798-4101-8a6a-a7f574ee693a","Type":"ContainerStarted","Data":"501ecbefbd40e7f91b0148575760a997fabc9039f716c2019e6c1ec4e8b15aa9"} Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.342318 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.343173 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" event={"ID":"7e4e41b5-dc12-4552-a5f2-28b01e76b84d","Type":"ContainerStarted","Data":"d47695b62eceb7385d35bb433c82c737c5a4440c5fbd72a154ff9e86e7c8bf7e"} Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.343341 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.359808 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" podStartSLOduration=1.5837783079999999 podStartE2EDuration="7.359790555s" podCreationTimestamp="2025-10-03 13:46:28 +0000 UTC" firstStartedPulling="2025-10-03 13:46:29.250917313 +0000 UTC m=+958.454260730" lastFinishedPulling="2025-10-03 13:46:35.02692955 +0000 UTC m=+964.230272977" observedRunningTime="2025-10-03 13:46:35.357403808 +0000 UTC m=+964.560747225" watchObservedRunningTime="2025-10-03 13:46:35.359790555 +0000 UTC m=+964.563133972" Oct 03 13:46:35 crc kubenswrapper[4959]: I1003 13:46:35.380895 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" podStartSLOduration=1.479000539 podStartE2EDuration="7.380879943s" podCreationTimestamp="2025-10-03 13:46:28 +0000 UTC" firstStartedPulling="2025-10-03 13:46:29.109236032 +0000 UTC m=+958.312579449" lastFinishedPulling="2025-10-03 13:46:35.011115436 +0000 UTC m=+964.214458853" observedRunningTime="2025-10-03 13:46:35.377177396 +0000 UTC m=+964.580520813" watchObservedRunningTime="2025-10-03 13:46:35.380879943 +0000 UTC m=+964.584223360" Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.044016 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.044088 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.044141 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.044804 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.044896 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a" gracePeriod=600 Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.350014 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a" exitCode=0 Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.350210 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a"} Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.350562 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694"} Oct 03 13:46:36 crc kubenswrapper[4959]: I1003 13:46:36.350595 4959 scope.go:117] "RemoveContainer" containerID="f0684f82c12b3d8e533e8e66f121835676e2b8cb10acf8d95681bff9ae422810" Oct 03 13:46:49 crc kubenswrapper[4959]: I1003 13:46:49.024301 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-98cf898b6-fmszp" Oct 03 13:47:08 crc kubenswrapper[4959]: I1003 13:47:08.555694 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-fd787bc66-gxclx" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.257727 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-c29td"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.258580 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.260471 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-svsvs" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.260479 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.261845 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6bnnt"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.268402 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.270392 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-c29td"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.271844 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.272943 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.338277 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-jcmdd"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.339307 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.342074 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kp69x" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.342082 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.342330 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.342878 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.355639 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-n269n"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.356441 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.362354 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.367130 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-n269n"] Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50b78443-7441-4240-ab34-0073a051380d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432578 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-sockets\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432610 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432634 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4klt6\" (UniqueName: \"kubernetes.io/projected/f55245be-65fc-43bc-b25c-04f7b660be38-kube-api-access-4klt6\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432817 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-conf\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.432969 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metallb-excludel2\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433027 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f55245be-65fc-43bc-b25c-04f7b660be38-frr-startup\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433102 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x8pr\" (UniqueName: \"kubernetes.io/projected/50b78443-7441-4240-ab34-0073a051380d-kube-api-access-7x8pr\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433151 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433208 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f55245be-65fc-43bc-b25c-04f7b660be38-metrics-certs\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433233 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-metrics\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-reloader\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.433342 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5brbr\" (UniqueName: \"kubernetes.io/projected/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-kube-api-access-5brbr\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534638 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534692 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4klt6\" (UniqueName: \"kubernetes.io/projected/f55245be-65fc-43bc-b25c-04f7b660be38-kube-api-access-4klt6\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534721 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-conf\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534764 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metallb-excludel2\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534796 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-cert\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534819 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f55245be-65fc-43bc-b25c-04f7b660be38-frr-startup\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x8pr\" (UniqueName: \"kubernetes.io/projected/50b78443-7441-4240-ab34-0073a051380d-kube-api-access-7x8pr\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534886 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2mzg\" (UniqueName: \"kubernetes.io/projected/a731fa03-adee-4bcf-bf90-b5c763dba9da-kube-api-access-x2mzg\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534909 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534933 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f55245be-65fc-43bc-b25c-04f7b660be38-metrics-certs\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: E1003 13:47:09.534946 4959 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 03 13:47:09 crc kubenswrapper[4959]: E1003 13:47:09.535073 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs podName:d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f nodeName:}" failed. No retries permitted until 2025-10-03 13:47:10.035047219 +0000 UTC m=+999.238390656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs") pod "speaker-jcmdd" (UID: "d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f") : secret "speaker-certs-secret" not found Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.534954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-metrics\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535231 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-reloader\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535281 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-metrics-certs\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5brbr\" (UniqueName: \"kubernetes.io/projected/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-kube-api-access-5brbr\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535350 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50b78443-7441-4240-ab34-0073a051380d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535436 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-sockets\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535742 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-metrics\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: E1003 13:47:09.535909 4959 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 13:47:09 crc kubenswrapper[4959]: E1003 13:47:09.535950 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist podName:d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f nodeName:}" failed. No retries permitted until 2025-10-03 13:47:10.035937581 +0000 UTC m=+999.239280998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist") pod "speaker-jcmdd" (UID: "d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f") : secret "metallb-memberlist" not found Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.535971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-sockets\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.536116 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-reloader\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.536487 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f55245be-65fc-43bc-b25c-04f7b660be38-frr-startup\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.536743 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f55245be-65fc-43bc-b25c-04f7b660be38-frr-conf\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.537608 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metallb-excludel2\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.542467 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50b78443-7441-4240-ab34-0073a051380d-cert\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.549212 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f55245be-65fc-43bc-b25c-04f7b660be38-metrics-certs\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.558535 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4klt6\" (UniqueName: \"kubernetes.io/projected/f55245be-65fc-43bc-b25c-04f7b660be38-kube-api-access-4klt6\") pod \"frr-k8s-6bnnt\" (UID: \"f55245be-65fc-43bc-b25c-04f7b660be38\") " pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.558663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x8pr\" (UniqueName: \"kubernetes.io/projected/50b78443-7441-4240-ab34-0073a051380d-kube-api-access-7x8pr\") pod \"frr-k8s-webhook-server-64bf5d555-c29td\" (UID: \"50b78443-7441-4240-ab34-0073a051380d\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.560582 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5brbr\" (UniqueName: \"kubernetes.io/projected/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-kube-api-access-5brbr\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.572917 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.582723 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.636670 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-cert\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.636724 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2mzg\" (UniqueName: \"kubernetes.io/projected/a731fa03-adee-4bcf-bf90-b5c763dba9da-kube-api-access-x2mzg\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.636764 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-metrics-certs\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.638513 4959 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.640822 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-metrics-certs\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.652081 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a731fa03-adee-4bcf-bf90-b5c763dba9da-cert\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.656163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2mzg\" (UniqueName: \"kubernetes.io/projected/a731fa03-adee-4bcf-bf90-b5c763dba9da-kube-api-access-x2mzg\") pod \"controller-68d546b9d8-n269n\" (UID: \"a731fa03-adee-4bcf-bf90-b5c763dba9da\") " pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.670338 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:09 crc kubenswrapper[4959]: I1003 13:47:09.986947 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-c29td"] Oct 03 13:47:09 crc kubenswrapper[4959]: W1003 13:47:09.995739 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50b78443_7441_4240_ab34_0073a051380d.slice/crio-ba81f06fa05efdc0f377a9cd3144f0ec0647d4bf360c600d6382bd5cdaac0103 WatchSource:0}: Error finding container ba81f06fa05efdc0f377a9cd3144f0ec0647d4bf360c600d6382bd5cdaac0103: Status 404 returned error can't find the container with id ba81f06fa05efdc0f377a9cd3144f0ec0647d4bf360c600d6382bd5cdaac0103 Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.042282 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.042360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:10 crc kubenswrapper[4959]: E1003 13:47:10.042464 4959 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 13:47:10 crc kubenswrapper[4959]: E1003 13:47:10.042506 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist podName:d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f nodeName:}" failed. No retries permitted until 2025-10-03 13:47:11.042492533 +0000 UTC m=+1000.245835950 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist") pod "speaker-jcmdd" (UID: "d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f") : secret "metallb-memberlist" not found Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.048477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-metrics-certs\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.078207 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-n269n"] Oct 03 13:47:10 crc kubenswrapper[4959]: W1003 13:47:10.080970 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda731fa03_adee_4bcf_bf90_b5c763dba9da.slice/crio-c72ab655b42ada8cc5fb9d5db7afb6e9792aa20aa9b8f444b6828f11d73eccdc WatchSource:0}: Error finding container c72ab655b42ada8cc5fb9d5db7afb6e9792aa20aa9b8f444b6828f11d73eccdc: Status 404 returned error can't find the container with id c72ab655b42ada8cc5fb9d5db7afb6e9792aa20aa9b8f444b6828f11d73eccdc Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.544215 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"62a55fbd4aaa206c10a71623a3ca056ed0f85bb47296d9898d8ff5bc0c26e483"} Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.545326 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" event={"ID":"50b78443-7441-4240-ab34-0073a051380d","Type":"ContainerStarted","Data":"ba81f06fa05efdc0f377a9cd3144f0ec0647d4bf360c600d6382bd5cdaac0103"} Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.546985 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-n269n" event={"ID":"a731fa03-adee-4bcf-bf90-b5c763dba9da","Type":"ContainerStarted","Data":"1341563bf6e2b25a027724a1fe651149473febe9f0ac59befaed1e6280c55922"} Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.547009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-n269n" event={"ID":"a731fa03-adee-4bcf-bf90-b5c763dba9da","Type":"ContainerStarted","Data":"4245def5aee108d19a19fd979fbc0e1db5a541b0cdb19b5f30278c0b55397b22"} Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.547018 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-n269n" event={"ID":"a731fa03-adee-4bcf-bf90-b5c763dba9da","Type":"ContainerStarted","Data":"c72ab655b42ada8cc5fb9d5db7afb6e9792aa20aa9b8f444b6828f11d73eccdc"} Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.547202 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:10 crc kubenswrapper[4959]: I1003 13:47:10.563156 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-n269n" podStartSLOduration=1.563137064 podStartE2EDuration="1.563137064s" podCreationTimestamp="2025-10-03 13:47:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:47:10.558503422 +0000 UTC m=+999.761846839" watchObservedRunningTime="2025-10-03 13:47:10.563137064 +0000 UTC m=+999.766480491" Oct 03 13:47:11 crc kubenswrapper[4959]: I1003 13:47:11.055584 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:11 crc kubenswrapper[4959]: I1003 13:47:11.064868 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f-memberlist\") pod \"speaker-jcmdd\" (UID: \"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f\") " pod="metallb-system/speaker-jcmdd" Oct 03 13:47:11 crc kubenswrapper[4959]: I1003 13:47:11.152179 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jcmdd" Oct 03 13:47:11 crc kubenswrapper[4959]: W1003 13:47:11.168770 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3cea5e4_7724_41bd_b3e6_e7ebbb7f3b7f.slice/crio-cf86df8112fb2469c66131132a0e08e280b57184a75a02505b9637cc3d68b2fe WatchSource:0}: Error finding container cf86df8112fb2469c66131132a0e08e280b57184a75a02505b9637cc3d68b2fe: Status 404 returned error can't find the container with id cf86df8112fb2469c66131132a0e08e280b57184a75a02505b9637cc3d68b2fe Oct 03 13:47:11 crc kubenswrapper[4959]: I1003 13:47:11.556084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jcmdd" event={"ID":"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f","Type":"ContainerStarted","Data":"d71f65a9e054eb714051af680a003aa2a19c5d84d21316516a06107572520e27"} Oct 03 13:47:11 crc kubenswrapper[4959]: I1003 13:47:11.556466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jcmdd" event={"ID":"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f","Type":"ContainerStarted","Data":"cf86df8112fb2469c66131132a0e08e280b57184a75a02505b9637cc3d68b2fe"} Oct 03 13:47:12 crc kubenswrapper[4959]: I1003 13:47:12.566672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jcmdd" event={"ID":"d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f","Type":"ContainerStarted","Data":"cb0e84bd98c876a7dc2bf41178500a1d43fe53b5faa9c960396005645649babc"} Oct 03 13:47:12 crc kubenswrapper[4959]: I1003 13:47:12.566802 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-jcmdd" Oct 03 13:47:12 crc kubenswrapper[4959]: I1003 13:47:12.586428 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-jcmdd" podStartSLOduration=3.58641154 podStartE2EDuration="3.58641154s" podCreationTimestamp="2025-10-03 13:47:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:47:12.582522886 +0000 UTC m=+1001.785866303" watchObservedRunningTime="2025-10-03 13:47:12.58641154 +0000 UTC m=+1001.789754957" Oct 03 13:47:17 crc kubenswrapper[4959]: I1003 13:47:17.604987 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" event={"ID":"50b78443-7441-4240-ab34-0073a051380d","Type":"ContainerStarted","Data":"7cbd891940dc4ffea0077a64dd139d5f22524bf2de40e147f79d2f5ae8641909"} Oct 03 13:47:17 crc kubenswrapper[4959]: I1003 13:47:17.605624 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:17 crc kubenswrapper[4959]: I1003 13:47:17.607026 4959 generic.go:334] "Generic (PLEG): container finished" podID="f55245be-65fc-43bc-b25c-04f7b660be38" containerID="73df7529e4866d04978676084dc6f5503a0cb79d33f5678b81a072d91ff6be23" exitCode=0 Oct 03 13:47:17 crc kubenswrapper[4959]: I1003 13:47:17.607061 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerDied","Data":"73df7529e4866d04978676084dc6f5503a0cb79d33f5678b81a072d91ff6be23"} Oct 03 13:47:17 crc kubenswrapper[4959]: I1003 13:47:17.622637 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" podStartSLOduration=1.282460739 podStartE2EDuration="8.622618559s" podCreationTimestamp="2025-10-03 13:47:09 +0000 UTC" firstStartedPulling="2025-10-03 13:47:09.998966104 +0000 UTC m=+999.202309521" lastFinishedPulling="2025-10-03 13:47:17.339123914 +0000 UTC m=+1006.542467341" observedRunningTime="2025-10-03 13:47:17.621234206 +0000 UTC m=+1006.824577633" watchObservedRunningTime="2025-10-03 13:47:17.622618559 +0000 UTC m=+1006.825961976" Oct 03 13:47:18 crc kubenswrapper[4959]: I1003 13:47:18.614002 4959 generic.go:334] "Generic (PLEG): container finished" podID="f55245be-65fc-43bc-b25c-04f7b660be38" containerID="77ca97dad0134492a40fb23f4568ca2005046eb5d350e950352adba2c3170660" exitCode=0 Oct 03 13:47:18 crc kubenswrapper[4959]: I1003 13:47:18.614066 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerDied","Data":"77ca97dad0134492a40fb23f4568ca2005046eb5d350e950352adba2c3170660"} Oct 03 13:47:19 crc kubenswrapper[4959]: I1003 13:47:19.621365 4959 generic.go:334] "Generic (PLEG): container finished" podID="f55245be-65fc-43bc-b25c-04f7b660be38" containerID="8ab3464bd7b93e1a1431da46f59ad6f46533aa373e149d8ca48e789f78975ff0" exitCode=0 Oct 03 13:47:19 crc kubenswrapper[4959]: I1003 13:47:19.621406 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerDied","Data":"8ab3464bd7b93e1a1431da46f59ad6f46533aa373e149d8ca48e789f78975ff0"} Oct 03 13:47:20 crc kubenswrapper[4959]: I1003 13:47:20.631584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"4d853226969fe38274a4a2413928606650ea5d29fe19723529660c74f2d3a6eb"} Oct 03 13:47:20 crc kubenswrapper[4959]: I1003 13:47:20.631832 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"0d6181f69b623b771c6bba91ff34201f799195756fdd7f8529213164302b99f3"} Oct 03 13:47:20 crc kubenswrapper[4959]: I1003 13:47:20.631842 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"ec54db4ce7c33a6003bd06a8de9cb49424441ba69aa84434165b86717333c5a8"} Oct 03 13:47:20 crc kubenswrapper[4959]: I1003 13:47:20.631850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"55fc45429ca9547b4d4b3c6f7bf909d9e1e7a97a3cbc4506dcf96e55df597acc"} Oct 03 13:47:20 crc kubenswrapper[4959]: I1003 13:47:20.631859 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"b8060ad770ef363d7f82dab7b53acf840867023a96d16208d187b0751bfff58c"} Oct 03 13:47:21 crc kubenswrapper[4959]: I1003 13:47:21.155730 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-jcmdd" Oct 03 13:47:21 crc kubenswrapper[4959]: I1003 13:47:21.642339 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6bnnt" event={"ID":"f55245be-65fc-43bc-b25c-04f7b660be38","Type":"ContainerStarted","Data":"0bc96437323a8ad4792582f275035ec8fd53b6d207c8aa6632453c84357639fd"} Oct 03 13:47:21 crc kubenswrapper[4959]: I1003 13:47:21.718782 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6bnnt" podStartSLOduration=5.237496913 podStartE2EDuration="12.718755185s" podCreationTimestamp="2025-10-03 13:47:09 +0000 UTC" firstStartedPulling="2025-10-03 13:47:09.823230767 +0000 UTC m=+999.026574184" lastFinishedPulling="2025-10-03 13:47:17.304489039 +0000 UTC m=+1006.507832456" observedRunningTime="2025-10-03 13:47:21.688456234 +0000 UTC m=+1010.891799651" watchObservedRunningTime="2025-10-03 13:47:21.718755185 +0000 UTC m=+1010.922098602" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.480152 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql"] Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.481205 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.492407 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.496056 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql"] Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.621867 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.622214 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfqtv\" (UniqueName: \"kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.622377 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.647715 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.724153 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.724270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfqtv\" (UniqueName: \"kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.724300 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.724648 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.724692 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.741384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfqtv\" (UniqueName: \"kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.795057 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:22 crc kubenswrapper[4959]: I1003 13:47:22.982450 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql"] Oct 03 13:47:23 crc kubenswrapper[4959]: I1003 13:47:23.655071 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerStarted","Data":"0a95a421c93f02256b0c8ae324f52af50395a4bd8c8235a71ae2d7c8794dd034"} Oct 03 13:47:23 crc kubenswrapper[4959]: I1003 13:47:23.655484 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerStarted","Data":"05ecd1182e5334dc0970b2e8de96a100b7befae5881b31d758d72ba4378509a8"} Oct 03 13:47:24 crc kubenswrapper[4959]: I1003 13:47:24.583267 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:24 crc kubenswrapper[4959]: I1003 13:47:24.620756 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:24 crc kubenswrapper[4959]: I1003 13:47:24.662899 4959 generic.go:334] "Generic (PLEG): container finished" podID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerID="0a95a421c93f02256b0c8ae324f52af50395a4bd8c8235a71ae2d7c8794dd034" exitCode=0 Oct 03 13:47:24 crc kubenswrapper[4959]: I1003 13:47:24.662950 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerDied","Data":"0a95a421c93f02256b0c8ae324f52af50395a4bd8c8235a71ae2d7c8794dd034"} Oct 03 13:47:29 crc kubenswrapper[4959]: I1003 13:47:29.588816 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6bnnt" Oct 03 13:47:29 crc kubenswrapper[4959]: I1003 13:47:29.606584 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-c29td" Oct 03 13:47:29 crc kubenswrapper[4959]: I1003 13:47:29.675452 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-n269n" Oct 03 13:47:29 crc kubenswrapper[4959]: I1003 13:47:29.693086 4959 generic.go:334] "Generic (PLEG): container finished" podID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerID="3a557acd87ac634838722b0087366c5646ef30d43f06b0e2ea3482771cec5c62" exitCode=0 Oct 03 13:47:29 crc kubenswrapper[4959]: I1003 13:47:29.700210 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerDied","Data":"3a557acd87ac634838722b0087366c5646ef30d43f06b0e2ea3482771cec5c62"} Oct 03 13:47:30 crc kubenswrapper[4959]: I1003 13:47:30.701829 4959 generic.go:334] "Generic (PLEG): container finished" podID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerID="1338b722d174975e41c835a0f8037415e4b4b6eca548103a5b9989055e5df529" exitCode=0 Oct 03 13:47:30 crc kubenswrapper[4959]: I1003 13:47:30.701882 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerDied","Data":"1338b722d174975e41c835a0f8037415e4b4b6eca548103a5b9989055e5df529"} Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.577268 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.725549 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" event={"ID":"34ed9cd3-5382-4ab8-98c0-cc5c1e305148","Type":"ContainerDied","Data":"05ecd1182e5334dc0970b2e8de96a100b7befae5881b31d758d72ba4378509a8"} Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.725590 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05ecd1182e5334dc0970b2e8de96a100b7befae5881b31d758d72ba4378509a8" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.725642 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.765120 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util\") pod \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.765272 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle\") pod \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.765673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfqtv\" (UniqueName: \"kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv\") pod \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\" (UID: \"34ed9cd3-5382-4ab8-98c0-cc5c1e305148\") " Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.766785 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle" (OuterVolumeSpecName: "bundle") pod "34ed9cd3-5382-4ab8-98c0-cc5c1e305148" (UID: "34ed9cd3-5382-4ab8-98c0-cc5c1e305148"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.772998 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv" (OuterVolumeSpecName: "kube-api-access-pfqtv") pod "34ed9cd3-5382-4ab8-98c0-cc5c1e305148" (UID: "34ed9cd3-5382-4ab8-98c0-cc5c1e305148"). InnerVolumeSpecName "kube-api-access-pfqtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.780807 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util" (OuterVolumeSpecName: "util") pod "34ed9cd3-5382-4ab8-98c0-cc5c1e305148" (UID: "34ed9cd3-5382-4ab8-98c0-cc5c1e305148"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.866847 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfqtv\" (UniqueName: \"kubernetes.io/projected/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-kube-api-access-pfqtv\") on node \"crc\" DevicePath \"\"" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.866885 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:47:32 crc kubenswrapper[4959]: I1003 13:47:32.866894 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/34ed9cd3-5382-4ab8-98c0-cc5c1e305148-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.144421 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw"] Oct 03 13:47:41 crc kubenswrapper[4959]: E1003 13:47:41.145244 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="extract" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.145261 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="extract" Oct 03 13:47:41 crc kubenswrapper[4959]: E1003 13:47:41.145273 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="util" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.145281 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="util" Oct 03 13:47:41 crc kubenswrapper[4959]: E1003 13:47:41.145303 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="pull" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.145312 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="pull" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.145431 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ed9cd3-5382-4ab8-98c0-cc5c1e305148" containerName="extract" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.145901 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.148784 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.149408 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.149406 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-lgnn7" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.164423 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw"] Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.318368 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gllqb\" (UniqueName: \"kubernetes.io/projected/cce26493-a548-4461-aaee-26b3df076fe0-kube-api-access-gllqb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-ffhkw\" (UID: \"cce26493-a548-4461-aaee-26b3df076fe0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.419669 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gllqb\" (UniqueName: \"kubernetes.io/projected/cce26493-a548-4461-aaee-26b3df076fe0-kube-api-access-gllqb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-ffhkw\" (UID: \"cce26493-a548-4461-aaee-26b3df076fe0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.440420 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gllqb\" (UniqueName: \"kubernetes.io/projected/cce26493-a548-4461-aaee-26b3df076fe0-kube-api-access-gllqb\") pod \"cert-manager-operator-controller-manager-57cd46d6d-ffhkw\" (UID: \"cce26493-a548-4461-aaee-26b3df076fe0\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.463292 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.659287 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw"] Oct 03 13:47:41 crc kubenswrapper[4959]: W1003 13:47:41.667923 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcce26493_a548_4461_aaee_26b3df076fe0.slice/crio-ea165878704c2e08491ade19f7c545f47a616089489a2bb47dbc683e60abadbb WatchSource:0}: Error finding container ea165878704c2e08491ade19f7c545f47a616089489a2bb47dbc683e60abadbb: Status 404 returned error can't find the container with id ea165878704c2e08491ade19f7c545f47a616089489a2bb47dbc683e60abadbb Oct 03 13:47:41 crc kubenswrapper[4959]: I1003 13:47:41.775435 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" event={"ID":"cce26493-a548-4461-aaee-26b3df076fe0","Type":"ContainerStarted","Data":"ea165878704c2e08491ade19f7c545f47a616089489a2bb47dbc683e60abadbb"} Oct 03 13:47:49 crc kubenswrapper[4959]: I1003 13:47:49.824701 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" event={"ID":"cce26493-a548-4461-aaee-26b3df076fe0","Type":"ContainerStarted","Data":"4b906fc8809e67c2ab8099e430fa72aca278952d2507183fc1b2092937c6b537"} Oct 03 13:47:49 crc kubenswrapper[4959]: I1003 13:47:49.842422 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-ffhkw" podStartSLOduration=1.6177410970000001 podStartE2EDuration="8.842405273s" podCreationTimestamp="2025-10-03 13:47:41 +0000 UTC" firstStartedPulling="2025-10-03 13:47:41.672013728 +0000 UTC m=+1030.875357145" lastFinishedPulling="2025-10-03 13:47:48.896677904 +0000 UTC m=+1038.100021321" observedRunningTime="2025-10-03 13:47:49.840548908 +0000 UTC m=+1039.043892345" watchObservedRunningTime="2025-10-03 13:47:49.842405273 +0000 UTC m=+1039.045748690" Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.900022 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rmd7h"] Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.900901 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.902836 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.903098 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.905764 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-pb4vk" Oct 03 13:47:51 crc kubenswrapper[4959]: I1003 13:47:51.912934 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rmd7h"] Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.051105 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.051263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hgd\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-kube-api-access-d7hgd\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.152381 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hgd\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-kube-api-access-d7hgd\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.152443 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.180750 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hgd\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-kube-api-access-d7hgd\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.182743 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4b67815b-67b5-40f5-b2e6-042e8910bf7f-bound-sa-token\") pod \"cert-manager-webhook-d969966f-rmd7h\" (UID: \"4b67815b-67b5-40f5-b2e6-042e8910bf7f\") " pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.214017 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.664705 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-rmd7h"] Oct 03 13:47:52 crc kubenswrapper[4959]: W1003 13:47:52.674238 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b67815b_67b5_40f5_b2e6_042e8910bf7f.slice/crio-3b1ec77964bad734a3c995057003163871831e3c40f288ed6be47fe3d7fc0863 WatchSource:0}: Error finding container 3b1ec77964bad734a3c995057003163871831e3c40f288ed6be47fe3d7fc0863: Status 404 returned error can't find the container with id 3b1ec77964bad734a3c995057003163871831e3c40f288ed6be47fe3d7fc0863 Oct 03 13:47:52 crc kubenswrapper[4959]: I1003 13:47:52.841610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" event={"ID":"4b67815b-67b5-40f5-b2e6-042e8910bf7f","Type":"ContainerStarted","Data":"3b1ec77964bad734a3c995057003163871831e3c40f288ed6be47fe3d7fc0863"} Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.823408 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p"] Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.824646 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.829480 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bchwj" Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.832179 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p"] Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.885435 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.886314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hdgp\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-kube-api-access-9hdgp\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.987183 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:54 crc kubenswrapper[4959]: I1003 13:47:54.987283 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hdgp\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-kube-api-access-9hdgp\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:55 crc kubenswrapper[4959]: I1003 13:47:55.011102 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hdgp\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-kube-api-access-9hdgp\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:55 crc kubenswrapper[4959]: I1003 13:47:55.019340 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/189b87c9-9c2b-4fdf-ad19-46346cfe3d5a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-hmw5p\" (UID: \"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:55 crc kubenswrapper[4959]: I1003 13:47:55.145752 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" Oct 03 13:47:55 crc kubenswrapper[4959]: I1003 13:47:55.345666 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p"] Oct 03 13:47:55 crc kubenswrapper[4959]: I1003 13:47:55.866965 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" event={"ID":"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a","Type":"ContainerStarted","Data":"36b294fc069c6e419f152e305af3a8d9b9cf2fc02fb334af30198aa6659c0bd2"} Oct 03 13:47:58 crc kubenswrapper[4959]: I1003 13:47:58.885784 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" event={"ID":"4b67815b-67b5-40f5-b2e6-042e8910bf7f","Type":"ContainerStarted","Data":"fed5dfcb74a0796f5e045c22397eca81e7edb60b59bc44033478f70b26c03d42"} Oct 03 13:47:58 crc kubenswrapper[4959]: I1003 13:47:58.886361 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:47:58 crc kubenswrapper[4959]: I1003 13:47:58.887444 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" event={"ID":"189b87c9-9c2b-4fdf-ad19-46346cfe3d5a","Type":"ContainerStarted","Data":"fcfa47e8e5c7e373d2cd6f590153e781b837666c7c4f06a6e93bfd97a165b39f"} Oct 03 13:47:58 crc kubenswrapper[4959]: I1003 13:47:58.901884 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" podStartSLOduration=2.2220779569999998 podStartE2EDuration="7.901864582s" podCreationTimestamp="2025-10-03 13:47:51 +0000 UTC" firstStartedPulling="2025-10-03 13:47:52.677727304 +0000 UTC m=+1041.881070721" lastFinishedPulling="2025-10-03 13:47:58.357513929 +0000 UTC m=+1047.560857346" observedRunningTime="2025-10-03 13:47:58.899146246 +0000 UTC m=+1048.102489683" watchObservedRunningTime="2025-10-03 13:47:58.901864582 +0000 UTC m=+1048.105208009" Oct 03 13:48:07 crc kubenswrapper[4959]: I1003 13:48:07.217648 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-rmd7h" Oct 03 13:48:07 crc kubenswrapper[4959]: I1003 13:48:07.232837 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-hmw5p" podStartSLOduration=10.233213426 podStartE2EDuration="13.232818678s" podCreationTimestamp="2025-10-03 13:47:54 +0000 UTC" firstStartedPulling="2025-10-03 13:47:55.357918367 +0000 UTC m=+1044.561261784" lastFinishedPulling="2025-10-03 13:47:58.357523619 +0000 UTC m=+1047.560867036" observedRunningTime="2025-10-03 13:47:58.917467288 +0000 UTC m=+1048.120810715" watchObservedRunningTime="2025-10-03 13:48:07.232818678 +0000 UTC m=+1056.436162095" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.007799 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-q5mrk"] Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.008890 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.012668 4959 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-q5b5x" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.031604 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-q5mrk"] Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.120970 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p62r\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-kube-api-access-7p62r\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.121045 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.222419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p62r\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-kube-api-access-7p62r\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.222486 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.239455 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.239612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p62r\" (UniqueName: \"kubernetes.io/projected/c1796b74-da6c-41f4-88e5-69bfec969799-kube-api-access-7p62r\") pod \"cert-manager-7d4cc89fcb-q5mrk\" (UID: \"c1796b74-da6c-41f4-88e5-69bfec969799\") " pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.327753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.748046 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-q5mrk"] Oct 03 13:48:11 crc kubenswrapper[4959]: W1003 13:48:11.751762 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1796b74_da6c_41f4_88e5_69bfec969799.slice/crio-b126a603a2a124d14a63a13fcbff1f61ca6907469a07ba16fa25b8936a7e6bf0 WatchSource:0}: Error finding container b126a603a2a124d14a63a13fcbff1f61ca6907469a07ba16fa25b8936a7e6bf0: Status 404 returned error can't find the container with id b126a603a2a124d14a63a13fcbff1f61ca6907469a07ba16fa25b8936a7e6bf0 Oct 03 13:48:11 crc kubenswrapper[4959]: I1003 13:48:11.973264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" event={"ID":"c1796b74-da6c-41f4-88e5-69bfec969799","Type":"ContainerStarted","Data":"b126a603a2a124d14a63a13fcbff1f61ca6907469a07ba16fa25b8936a7e6bf0"} Oct 03 13:48:12 crc kubenswrapper[4959]: I1003 13:48:12.980216 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" event={"ID":"c1796b74-da6c-41f4-88e5-69bfec969799","Type":"ContainerStarted","Data":"887d10b66e9418c0002d26a1d38dca87ccd6430706abc90d3550704b9c1e7342"} Oct 03 13:48:13 crc kubenswrapper[4959]: I1003 13:48:13.000069 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-q5mrk" podStartSLOduration=3.000045861 podStartE2EDuration="3.000045861s" podCreationTimestamp="2025-10-03 13:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:48:12.99958068 +0000 UTC m=+1062.202924107" watchObservedRunningTime="2025-10-03 13:48:13.000045861 +0000 UTC m=+1062.203389308" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.517261 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.518728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.521728 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.521764 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.521905 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-rtsgv" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.535867 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.662654 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bpp\" (UniqueName: \"kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp\") pod \"openstack-operator-index-5vrm2\" (UID: \"5a2f4162-932a-413b-8cad-fefed97f8c4c\") " pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.764364 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7bpp\" (UniqueName: \"kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp\") pod \"openstack-operator-index-5vrm2\" (UID: \"5a2f4162-932a-413b-8cad-fefed97f8c4c\") " pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.783456 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7bpp\" (UniqueName: \"kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp\") pod \"openstack-operator-index-5vrm2\" (UID: \"5a2f4162-932a-413b-8cad-fefed97f8c4c\") " pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:20 crc kubenswrapper[4959]: I1003 13:48:20.845997 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:21 crc kubenswrapper[4959]: I1003 13:48:21.244163 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:22 crc kubenswrapper[4959]: I1003 13:48:22.035998 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5vrm2" event={"ID":"5a2f4162-932a-413b-8cad-fefed97f8c4c","Type":"ContainerStarted","Data":"90620bc053a810352ae66540d5f3e9ad7c670982ee1a305d09c61f3678592fa6"} Oct 03 13:48:23 crc kubenswrapper[4959]: I1003 13:48:23.695533 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.302012 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hj995"] Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.302735 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.311883 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hj995"] Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.410601 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfhnq\" (UniqueName: \"kubernetes.io/projected/39879f5c-73d5-43cf-be16-aadbe112f51b-kube-api-access-sfhnq\") pod \"openstack-operator-index-hj995\" (UID: \"39879f5c-73d5-43cf-be16-aadbe112f51b\") " pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.512490 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfhnq\" (UniqueName: \"kubernetes.io/projected/39879f5c-73d5-43cf-be16-aadbe112f51b-kube-api-access-sfhnq\") pod \"openstack-operator-index-hj995\" (UID: \"39879f5c-73d5-43cf-be16-aadbe112f51b\") " pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.530162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfhnq\" (UniqueName: \"kubernetes.io/projected/39879f5c-73d5-43cf-be16-aadbe112f51b-kube-api-access-sfhnq\") pod \"openstack-operator-index-hj995\" (UID: \"39879f5c-73d5-43cf-be16-aadbe112f51b\") " pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:24 crc kubenswrapper[4959]: I1003 13:48:24.632285 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:26 crc kubenswrapper[4959]: I1003 13:48:26.058398 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hj995"] Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.070821 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-5vrm2" podUID="5a2f4162-932a-413b-8cad-fefed97f8c4c" containerName="registry-server" containerID="cri-o://fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9" gracePeriod=2 Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.071168 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5vrm2" event={"ID":"5a2f4162-932a-413b-8cad-fefed97f8c4c","Type":"ContainerStarted","Data":"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9"} Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.074587 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hj995" event={"ID":"39879f5c-73d5-43cf-be16-aadbe112f51b","Type":"ContainerStarted","Data":"716810f370546c45d90b06eba21be68728725d45f7db52b6b93b74ecedb99987"} Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.074695 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hj995" event={"ID":"39879f5c-73d5-43cf-be16-aadbe112f51b","Type":"ContainerStarted","Data":"047158834b7845cc9927c3712f66c96ab82cec98aed488888c544f71036cdbe1"} Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.105680 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hj995" podStartSLOduration=2.944118273 podStartE2EDuration="3.105661778s" podCreationTimestamp="2025-10-03 13:48:24 +0000 UTC" firstStartedPulling="2025-10-03 13:48:26.073880384 +0000 UTC m=+1075.277223801" lastFinishedPulling="2025-10-03 13:48:26.235423869 +0000 UTC m=+1075.438767306" observedRunningTime="2025-10-03 13:48:27.104940611 +0000 UTC m=+1076.308284028" watchObservedRunningTime="2025-10-03 13:48:27.105661778 +0000 UTC m=+1076.309005195" Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.107289 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5vrm2" podStartSLOduration=2.387133358 podStartE2EDuration="7.107282858s" podCreationTimestamp="2025-10-03 13:48:20 +0000 UTC" firstStartedPulling="2025-10-03 13:48:21.250415264 +0000 UTC m=+1070.453758681" lastFinishedPulling="2025-10-03 13:48:25.970564764 +0000 UTC m=+1075.173908181" observedRunningTime="2025-10-03 13:48:27.090997835 +0000 UTC m=+1076.294341262" watchObservedRunningTime="2025-10-03 13:48:27.107282858 +0000 UTC m=+1076.310626275" Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.478894 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.652293 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7bpp\" (UniqueName: \"kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp\") pod \"5a2f4162-932a-413b-8cad-fefed97f8c4c\" (UID: \"5a2f4162-932a-413b-8cad-fefed97f8c4c\") " Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.657745 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp" (OuterVolumeSpecName: "kube-api-access-v7bpp") pod "5a2f4162-932a-413b-8cad-fefed97f8c4c" (UID: "5a2f4162-932a-413b-8cad-fefed97f8c4c"). InnerVolumeSpecName "kube-api-access-v7bpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:48:27 crc kubenswrapper[4959]: I1003 13:48:27.753735 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7bpp\" (UniqueName: \"kubernetes.io/projected/5a2f4162-932a-413b-8cad-fefed97f8c4c-kube-api-access-v7bpp\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.081427 4959 generic.go:334] "Generic (PLEG): container finished" podID="5a2f4162-932a-413b-8cad-fefed97f8c4c" containerID="fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9" exitCode=0 Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.081504 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5vrm2" Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.081500 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5vrm2" event={"ID":"5a2f4162-932a-413b-8cad-fefed97f8c4c","Type":"ContainerDied","Data":"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9"} Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.081602 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5vrm2" event={"ID":"5a2f4162-932a-413b-8cad-fefed97f8c4c","Type":"ContainerDied","Data":"90620bc053a810352ae66540d5f3e9ad7c670982ee1a305d09c61f3678592fa6"} Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.081645 4959 scope.go:117] "RemoveContainer" containerID="fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9" Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.104422 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.104748 4959 scope.go:117] "RemoveContainer" containerID="fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9" Oct 03 13:48:28 crc kubenswrapper[4959]: E1003 13:48:28.105215 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9\": container with ID starting with fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9 not found: ID does not exist" containerID="fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9" Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.105258 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9"} err="failed to get container status \"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9\": rpc error: code = NotFound desc = could not find container \"fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9\": container with ID starting with fc7fa1e7b5a45c66f3a55c9769e89242032ccfe690339932813d9c0473b4f5c9 not found: ID does not exist" Oct 03 13:48:28 crc kubenswrapper[4959]: I1003 13:48:28.109489 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-5vrm2"] Oct 03 13:48:29 crc kubenswrapper[4959]: I1003 13:48:29.692341 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a2f4162-932a-413b-8cad-fefed97f8c4c" path="/var/lib/kubelet/pods/5a2f4162-932a-413b-8cad-fefed97f8c4c/volumes" Oct 03 13:48:34 crc kubenswrapper[4959]: I1003 13:48:34.633428 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:34 crc kubenswrapper[4959]: I1003 13:48:34.633811 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:34 crc kubenswrapper[4959]: I1003 13:48:34.666892 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:35 crc kubenswrapper[4959]: I1003 13:48:35.155820 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hj995" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.044929 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.046368 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.361509 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z"] Oct 03 13:48:36 crc kubenswrapper[4959]: E1003 13:48:36.361903 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2f4162-932a-413b-8cad-fefed97f8c4c" containerName="registry-server" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.361936 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2f4162-932a-413b-8cad-fefed97f8c4c" containerName="registry-server" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.362159 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2f4162-932a-413b-8cad-fefed97f8c4c" containerName="registry-server" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.363750 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.367304 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vn2m2" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.369737 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z"] Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.468265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lld7v\" (UniqueName: \"kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.468362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.468421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.569787 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lld7v\" (UniqueName: \"kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.569863 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.569929 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.570440 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.570771 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.603802 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lld7v\" (UniqueName: \"kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v\") pod \"4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.685778 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:36 crc kubenswrapper[4959]: I1003 13:48:36.885759 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z"] Oct 03 13:48:37 crc kubenswrapper[4959]: I1003 13:48:37.144484 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerStarted","Data":"6d743a4dd02a7ab5a9e8e0900d600e2df35c9777bcb4c39d052dd1358cbbbbd1"} Oct 03 13:48:37 crc kubenswrapper[4959]: I1003 13:48:37.144529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerStarted","Data":"c5c261af146836fe3a87fea750381be6a48605b8ca1f1ebf7080236b1a08bdcc"} Oct 03 13:48:38 crc kubenswrapper[4959]: I1003 13:48:38.155649 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerID="6d743a4dd02a7ab5a9e8e0900d600e2df35c9777bcb4c39d052dd1358cbbbbd1" exitCode=0 Oct 03 13:48:38 crc kubenswrapper[4959]: I1003 13:48:38.155714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerDied","Data":"6d743a4dd02a7ab5a9e8e0900d600e2df35c9777bcb4c39d052dd1358cbbbbd1"} Oct 03 13:48:40 crc kubenswrapper[4959]: I1003 13:48:40.183699 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerID="eb4964c05e1f7c5414cea18c94ec72fe0637b2d9b08e8a7777258e2d81e9b3bf" exitCode=0 Oct 03 13:48:40 crc kubenswrapper[4959]: I1003 13:48:40.183818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerDied","Data":"eb4964c05e1f7c5414cea18c94ec72fe0637b2d9b08e8a7777258e2d81e9b3bf"} Oct 03 13:48:41 crc kubenswrapper[4959]: I1003 13:48:41.194392 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerID="36fd7e4056f82a125c021af0f49b11f62532e19458d8b9eba3727084d1153f92" exitCode=0 Oct 03 13:48:41 crc kubenswrapper[4959]: I1003 13:48:41.194456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerDied","Data":"36fd7e4056f82a125c021af0f49b11f62532e19458d8b9eba3727084d1153f92"} Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.496515 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.652440 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util\") pod \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.652544 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lld7v\" (UniqueName: \"kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v\") pod \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.652637 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle\") pod \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\" (UID: \"ea1ffdb4-f9b4-42aa-b757-f04040023dc6\") " Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.654899 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle" (OuterVolumeSpecName: "bundle") pod "ea1ffdb4-f9b4-42aa-b757-f04040023dc6" (UID: "ea1ffdb4-f9b4-42aa-b757-f04040023dc6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.658860 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v" (OuterVolumeSpecName: "kube-api-access-lld7v") pod "ea1ffdb4-f9b4-42aa-b757-f04040023dc6" (UID: "ea1ffdb4-f9b4-42aa-b757-f04040023dc6"). InnerVolumeSpecName "kube-api-access-lld7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.754235 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lld7v\" (UniqueName: \"kubernetes.io/projected/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-kube-api-access-lld7v\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.754263 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.939434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util" (OuterVolumeSpecName: "util") pod "ea1ffdb4-f9b4-42aa-b757-f04040023dc6" (UID: "ea1ffdb4-f9b4-42aa-b757-f04040023dc6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:48:42 crc kubenswrapper[4959]: I1003 13:48:42.956858 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ea1ffdb4-f9b4-42aa-b757-f04040023dc6-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:43 crc kubenswrapper[4959]: I1003 13:48:43.212813 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" Oct 03 13:48:43 crc kubenswrapper[4959]: I1003 13:48:43.212803 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z" event={"ID":"ea1ffdb4-f9b4-42aa-b757-f04040023dc6","Type":"ContainerDied","Data":"c5c261af146836fe3a87fea750381be6a48605b8ca1f1ebf7080236b1a08bdcc"} Oct 03 13:48:43 crc kubenswrapper[4959]: I1003 13:48:43.212966 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c261af146836fe3a87fea750381be6a48605b8ca1f1ebf7080236b1a08bdcc" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.082621 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8"] Oct 03 13:48:49 crc kubenswrapper[4959]: E1003 13:48:49.083769 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="extract" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.083789 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="extract" Oct 03 13:48:49 crc kubenswrapper[4959]: E1003 13:48:49.083817 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="pull" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.083825 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="pull" Oct 03 13:48:49 crc kubenswrapper[4959]: E1003 13:48:49.083858 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="util" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.083865 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="util" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.084112 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1ffdb4-f9b4-42aa-b757-f04040023dc6" containerName="extract" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.086461 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.111677 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gndmw" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.123950 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8"] Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.243715 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-559sk\" (UniqueName: \"kubernetes.io/projected/dae17dc0-d946-464d-a655-93ad2ba51862-kube-api-access-559sk\") pod \"openstack-operator-controller-operator-5b7969687c-mqdr8\" (UID: \"dae17dc0-d946-464d-a655-93ad2ba51862\") " pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.344754 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-559sk\" (UniqueName: \"kubernetes.io/projected/dae17dc0-d946-464d-a655-93ad2ba51862-kube-api-access-559sk\") pod \"openstack-operator-controller-operator-5b7969687c-mqdr8\" (UID: \"dae17dc0-d946-464d-a655-93ad2ba51862\") " pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.366477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-559sk\" (UniqueName: \"kubernetes.io/projected/dae17dc0-d946-464d-a655-93ad2ba51862-kube-api-access-559sk\") pod \"openstack-operator-controller-operator-5b7969687c-mqdr8\" (UID: \"dae17dc0-d946-464d-a655-93ad2ba51862\") " pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.416766 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:48:49 crc kubenswrapper[4959]: I1003 13:48:49.611753 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8"] Oct 03 13:48:50 crc kubenswrapper[4959]: I1003 13:48:50.263959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" event={"ID":"dae17dc0-d946-464d-a655-93ad2ba51862","Type":"ContainerStarted","Data":"e0f5bea1b4d6258fbb094906a569f420a33cffc5e0acbdf560f796c6bb7660ac"} Oct 03 13:48:55 crc kubenswrapper[4959]: I1003 13:48:55.297770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" event={"ID":"dae17dc0-d946-464d-a655-93ad2ba51862","Type":"ContainerStarted","Data":"1337b70938bb100d849b1c349b3e1250027852a3005012f754e1a3faf9565f11"} Oct 03 13:49:03 crc kubenswrapper[4959]: I1003 13:49:03.355775 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" event={"ID":"dae17dc0-d946-464d-a655-93ad2ba51862","Type":"ContainerStarted","Data":"ad64cfcda97f09c84c70803a4fcf0a8ae8e80550eae5049e31bc43402f780fb6"} Oct 03 13:49:03 crc kubenswrapper[4959]: I1003 13:49:03.356316 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:49:03 crc kubenswrapper[4959]: I1003 13:49:03.358361 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" Oct 03 13:49:03 crc kubenswrapper[4959]: I1003 13:49:03.386568 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5b7969687c-mqdr8" podStartSLOduration=1.159850535 podStartE2EDuration="14.386546935s" podCreationTimestamp="2025-10-03 13:48:49 +0000 UTC" firstStartedPulling="2025-10-03 13:48:49.622218504 +0000 UTC m=+1098.825561921" lastFinishedPulling="2025-10-03 13:49:02.848914904 +0000 UTC m=+1112.052258321" observedRunningTime="2025-10-03 13:49:03.383098811 +0000 UTC m=+1112.586442248" watchObservedRunningTime="2025-10-03 13:49:03.386546935 +0000 UTC m=+1112.589890342" Oct 03 13:49:06 crc kubenswrapper[4959]: I1003 13:49:06.044375 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:49:06 crc kubenswrapper[4959]: I1003 13:49:06.044658 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.117468 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.118870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.122597 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hmtfz" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.126672 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.127675 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.132509 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-625kw" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.136485 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.139932 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.142086 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cl9m6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.160932 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.183421 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.192055 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.193389 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.196212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqhd6\" (UniqueName: \"kubernetes.io/projected/b991eb57-3139-462c-921b-72525e281d25-kube-api-access-vqhd6\") pod \"cinder-operator-controller-manager-79d68d6c85-dsq2k\" (UID: \"b991eb57-3139-462c-921b-72525e281d25\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.196325 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzrjj\" (UniqueName: \"kubernetes.io/projected/1848147c-d228-4d3c-bace-7628a688438d-kube-api-access-gzrjj\") pod \"barbican-operator-controller-manager-6c675fb79f-n78z6\" (UID: \"1848147c-d228-4d3c-bace-7628a688438d\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.196353 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kljtq\" (UniqueName: \"kubernetes.io/projected/f1efc6bb-e4c4-437f-8a29-8b587327652b-kube-api-access-kljtq\") pod \"designate-operator-controller-manager-75dfd9b554-5bgvs\" (UID: \"f1efc6bb-e4c4-437f-8a29-8b587327652b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.204019 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-2tpsr" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.227351 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.236911 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fdhll"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.243471 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.243941 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fdhll"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.243987 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.250350 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-pdrjj" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.255135 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.256021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.257765 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-vh2d6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.274180 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.284762 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.288063 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.292630 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.292909 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-56qz8" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297180 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xvsf\" (UniqueName: \"kubernetes.io/projected/c5519a95-6c42-4bc5-be67-626e8878421c-kube-api-access-5xvsf\") pod \"glance-operator-controller-manager-846dff85b5-zpl6z\" (UID: \"c5519a95-6c42-4bc5-be67-626e8878421c\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297375 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzrjj\" (UniqueName: \"kubernetes.io/projected/1848147c-d228-4d3c-bace-7628a688438d-kube-api-access-gzrjj\") pod \"barbican-operator-controller-manager-6c675fb79f-n78z6\" (UID: \"1848147c-d228-4d3c-bace-7628a688438d\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297440 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kljtq\" (UniqueName: \"kubernetes.io/projected/f1efc6bb-e4c4-437f-8a29-8b587327652b-kube-api-access-kljtq\") pod \"designate-operator-controller-manager-75dfd9b554-5bgvs\" (UID: \"f1efc6bb-e4c4-437f-8a29-8b587327652b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297479 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqhd6\" (UniqueName: \"kubernetes.io/projected/b991eb57-3139-462c-921b-72525e281d25-kube-api-access-vqhd6\") pod \"cinder-operator-controller-manager-79d68d6c85-dsq2k\" (UID: \"b991eb57-3139-462c-921b-72525e281d25\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297766 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q5zj\" (UniqueName: \"kubernetes.io/projected/a1ef21c1-ce09-47cc-b1c6-a7e8916431c6-kube-api-access-7q5zj\") pod \"horizon-operator-controller-manager-6769b867d9-crg46\" (UID: \"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.297815 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rkq5\" (UniqueName: \"kubernetes.io/projected/84bc4832-d055-4c00-955d-72e589902ea7-kube-api-access-6rkq5\") pod \"heat-operator-controller-manager-599898f689-fdhll\" (UID: \"84bc4832-d055-4c00-955d-72e589902ea7\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.310877 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.323362 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.324323 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.341154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzrjj\" (UniqueName: \"kubernetes.io/projected/1848147c-d228-4d3c-bace-7628a688438d-kube-api-access-gzrjj\") pod \"barbican-operator-controller-manager-6c675fb79f-n78z6\" (UID: \"1848147c-d228-4d3c-bace-7628a688438d\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.344946 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.345979 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.350862 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kljtq\" (UniqueName: \"kubernetes.io/projected/f1efc6bb-e4c4-437f-8a29-8b587327652b-kube-api-access-kljtq\") pod \"designate-operator-controller-manager-75dfd9b554-5bgvs\" (UID: \"f1efc6bb-e4c4-437f-8a29-8b587327652b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.351328 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2d8gc" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.351893 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-kqmfr" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.353424 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.354709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.361245 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.366096 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqhd6\" (UniqueName: \"kubernetes.io/projected/b991eb57-3139-462c-921b-72525e281d25-kube-api-access-vqhd6\") pod \"cinder-operator-controller-manager-79d68d6c85-dsq2k\" (UID: \"b991eb57-3139-462c-921b-72525e281d25\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.380214 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-h4f5t" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.380338 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.381386 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.383018 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.386338 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-gqvd7" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.394410 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400226 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xvsf\" (UniqueName: \"kubernetes.io/projected/c5519a95-6c42-4bc5-be67-626e8878421c-kube-api-access-5xvsf\") pod \"glance-operator-controller-manager-846dff85b5-zpl6z\" (UID: \"c5519a95-6c42-4bc5-be67-626e8878421c\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh5xx\" (UniqueName: \"kubernetes.io/projected/111756cf-9060-4c4b-878b-4ac998f89786-kube-api-access-mh5xx\") pod \"ironic-operator-controller-manager-84bc9db6cc-9ptc9\" (UID: \"111756cf-9060-4c4b-878b-4ac998f89786\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400431 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbwrt\" (UniqueName: \"kubernetes.io/projected/37c195cb-ee19-4040-ae6a-388685fed18d-kube-api-access-hbwrt\") pod \"keystone-operator-controller-manager-7f55849f88-w96jd\" (UID: \"37c195cb-ee19-4040-ae6a-388685fed18d\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400471 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zfxs\" (UniqueName: \"kubernetes.io/projected/0f49a60f-299a-4568-81ca-687eb8d439d3-kube-api-access-9zfxs\") pod \"manila-operator-controller-manager-6fd6854b49-55vd5\" (UID: \"0f49a60f-299a-4568-81ca-687eb8d439d3\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q5zj\" (UniqueName: \"kubernetes.io/projected/a1ef21c1-ce09-47cc-b1c6-a7e8916431c6-kube-api-access-7q5zj\") pod \"horizon-operator-controller-manager-6769b867d9-crg46\" (UID: \"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rkq5\" (UniqueName: \"kubernetes.io/projected/84bc4832-d055-4c00-955d-72e589902ea7-kube-api-access-6rkq5\") pod \"heat-operator-controller-manager-599898f689-fdhll\" (UID: \"84bc4832-d055-4c00-955d-72e589902ea7\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.400556 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sclfq\" (UniqueName: \"kubernetes.io/projected/e06190ac-1517-453b-a038-8437d0c77ee1-kube-api-access-sclfq\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.422322 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.433870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xvsf\" (UniqueName: \"kubernetes.io/projected/c5519a95-6c42-4bc5-be67-626e8878421c-kube-api-access-5xvsf\") pod \"glance-operator-controller-manager-846dff85b5-zpl6z\" (UID: \"c5519a95-6c42-4bc5-be67-626e8878421c\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.436840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q5zj\" (UniqueName: \"kubernetes.io/projected/a1ef21c1-ce09-47cc-b1c6-a7e8916431c6-kube-api-access-7q5zj\") pod \"horizon-operator-controller-manager-6769b867d9-crg46\" (UID: \"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.438710 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.445255 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.446295 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.469093 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rkq5\" (UniqueName: \"kubernetes.io/projected/84bc4832-d055-4c00-955d-72e589902ea7-kube-api-access-6rkq5\") pod \"heat-operator-controller-manager-599898f689-fdhll\" (UID: \"84bc4832-d055-4c00-955d-72e589902ea7\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.470004 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2xz6k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.470396 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.482847 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.489409 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.492016 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.493365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.494641 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fnkkd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.510824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zfxs\" (UniqueName: \"kubernetes.io/projected/0f49a60f-299a-4568-81ca-687eb8d439d3-kube-api-access-9zfxs\") pod \"manila-operator-controller-manager-6fd6854b49-55vd5\" (UID: \"0f49a60f-299a-4568-81ca-687eb8d439d3\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.510882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sclfq\" (UniqueName: \"kubernetes.io/projected/e06190ac-1517-453b-a038-8437d0c77ee1-kube-api-access-sclfq\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.510930 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.510983 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wctkw\" (UniqueName: \"kubernetes.io/projected/6aa2a2f3-7729-43a4-8564-c74ef4ddf86d-kube-api-access-wctkw\") pod \"neutron-operator-controller-manager-6574bf987d-h29xd\" (UID: \"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.511010 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh5xx\" (UniqueName: \"kubernetes.io/projected/111756cf-9060-4c4b-878b-4ac998f89786-kube-api-access-mh5xx\") pod \"ironic-operator-controller-manager-84bc9db6cc-9ptc9\" (UID: \"111756cf-9060-4c4b-878b-4ac998f89786\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.511060 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrs2h\" (UniqueName: \"kubernetes.io/projected/13a6dcff-d726-469a-9807-0ff2d76a7713-kube-api-access-jrs2h\") pod \"mariadb-operator-controller-manager-5c468bf4d4-qkmbn\" (UID: \"13a6dcff-d726-469a-9807-0ff2d76a7713\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.511087 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbwrt\" (UniqueName: \"kubernetes.io/projected/37c195cb-ee19-4040-ae6a-388685fed18d-kube-api-access-hbwrt\") pod \"keystone-operator-controller-manager-7f55849f88-w96jd\" (UID: \"37c195cb-ee19-4040-ae6a-388685fed18d\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:21 crc kubenswrapper[4959]: E1003 13:49:21.511754 4959 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 03 13:49:21 crc kubenswrapper[4959]: E1003 13:49:21.511812 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert podName:e06190ac-1517-453b-a038-8437d0c77ee1 nodeName:}" failed. No retries permitted until 2025-10-03 13:49:22.011792715 +0000 UTC m=+1131.215136132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert") pod "infra-operator-controller-manager-5fbf469cd7-cg44n" (UID: "e06190ac-1517-453b-a038-8437d0c77ee1") : secret "infra-operator-webhook-server-cert" not found Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.516911 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.527291 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.528331 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.531016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbwrt\" (UniqueName: \"kubernetes.io/projected/37c195cb-ee19-4040-ae6a-388685fed18d-kube-api-access-hbwrt\") pod \"keystone-operator-controller-manager-7f55849f88-w96jd\" (UID: \"37c195cb-ee19-4040-ae6a-388685fed18d\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.531861 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qthn9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.552859 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh5xx\" (UniqueName: \"kubernetes.io/projected/111756cf-9060-4c4b-878b-4ac998f89786-kube-api-access-mh5xx\") pod \"ironic-operator-controller-manager-84bc9db6cc-9ptc9\" (UID: \"111756cf-9060-4c4b-878b-4ac998f89786\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.553644 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sclfq\" (UniqueName: \"kubernetes.io/projected/e06190ac-1517-453b-a038-8437d0c77ee1-kube-api-access-sclfq\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.557267 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.561827 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zfxs\" (UniqueName: \"kubernetes.io/projected/0f49a60f-299a-4568-81ca-687eb8d439d3-kube-api-access-9zfxs\") pod \"manila-operator-controller-manager-6fd6854b49-55vd5\" (UID: \"0f49a60f-299a-4568-81ca-687eb8d439d3\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.575148 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.586946 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.592646 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.596155 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.596399 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.625869 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.627960 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-kthls" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.637746 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.640304 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-w7zlf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.649435 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.651850 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.653727 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wctkw\" (UniqueName: \"kubernetes.io/projected/6aa2a2f3-7729-43a4-8564-c74ef4ddf86d-kube-api-access-wctkw\") pod \"neutron-operator-controller-manager-6574bf987d-h29xd\" (UID: \"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.653908 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhghl\" (UniqueName: \"kubernetes.io/projected/469d9704-d797-4ab2-9e51-b4a2a62942b7-kube-api-access-lhghl\") pod \"nova-operator-controller-manager-555c7456bd-ssfjs\" (UID: \"469d9704-d797-4ab2-9e51-b4a2a62942b7\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.653966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrs2h\" (UniqueName: \"kubernetes.io/projected/13a6dcff-d726-469a-9807-0ff2d76a7713-kube-api-access-jrs2h\") pod \"mariadb-operator-controller-manager-5c468bf4d4-qkmbn\" (UID: \"13a6dcff-d726-469a-9807-0ff2d76a7713\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.654024 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlvxt\" (UniqueName: \"kubernetes.io/projected/ce1839e2-3240-40dd-9145-867852c68de2-kube-api-access-dlvxt\") pod \"octavia-operator-controller-manager-59d6cfdf45-rfvdf\" (UID: \"ce1839e2-3240-40dd-9145-867852c68de2\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.655855 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.665784 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.667175 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.669785 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.694972 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-kcfwg" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.695524 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-jfpjn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.719483 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.722532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrs2h\" (UniqueName: \"kubernetes.io/projected/13a6dcff-d726-469a-9807-0ff2d76a7713-kube-api-access-jrs2h\") pod \"mariadb-operator-controller-manager-5c468bf4d4-qkmbn\" (UID: \"13a6dcff-d726-469a-9807-0ff2d76a7713\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.725715 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wctkw\" (UniqueName: \"kubernetes.io/projected/6aa2a2f3-7729-43a4-8564-c74ef4ddf86d-kube-api-access-wctkw\") pod \"neutron-operator-controller-manager-6574bf987d-h29xd\" (UID: \"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.729348 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.735827 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755453 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755526 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwxv\" (UniqueName: \"kubernetes.io/projected/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-kube-api-access-mvwxv\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755564 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h42ml\" (UniqueName: \"kubernetes.io/projected/b4d484dc-fd46-432a-963d-049705db3803-kube-api-access-h42ml\") pod \"swift-operator-controller-manager-6859f9b676-rntmf\" (UID: \"b4d484dc-fd46-432a-963d-049705db3803\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755588 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjps4\" (UniqueName: \"kubernetes.io/projected/714d66a0-6e2e-4f57-8b39-39fff761851f-kube-api-access-mjps4\") pod \"placement-operator-controller-manager-7d8bb7f44c-rjzvl\" (UID: \"714d66a0-6e2e-4f57-8b39-39fff761851f\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755619 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhghl\" (UniqueName: \"kubernetes.io/projected/469d9704-d797-4ab2-9e51-b4a2a62942b7-kube-api-access-lhghl\") pod \"nova-operator-controller-manager-555c7456bd-ssfjs\" (UID: \"469d9704-d797-4ab2-9e51-b4a2a62942b7\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755653 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlvg\" (UniqueName: \"kubernetes.io/projected/6aef6e41-8436-4b85-8bf8-befc6314fee0-kube-api-access-fwlvg\") pod \"ovn-operator-controller-manager-688db7b6c7-r8w7q\" (UID: \"6aef6e41-8436-4b85-8bf8-befc6314fee0\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.755696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlvxt\" (UniqueName: \"kubernetes.io/projected/ce1839e2-3240-40dd-9145-867852c68de2-kube-api-access-dlvxt\") pod \"octavia-operator-controller-manager-59d6cfdf45-rfvdf\" (UID: \"ce1839e2-3240-40dd-9145-867852c68de2\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.770305 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.770353 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.783499 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.785673 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.789999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhghl\" (UniqueName: \"kubernetes.io/projected/469d9704-d797-4ab2-9e51-b4a2a62942b7-kube-api-access-lhghl\") pod \"nova-operator-controller-manager-555c7456bd-ssfjs\" (UID: \"469d9704-d797-4ab2-9e51-b4a2a62942b7\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.805347 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlvxt\" (UniqueName: \"kubernetes.io/projected/ce1839e2-3240-40dd-9145-867852c68de2-kube-api-access-dlvxt\") pod \"octavia-operator-controller-manager-59d6cfdf45-rfvdf\" (UID: \"ce1839e2-3240-40dd-9145-867852c68de2\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.805458 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-g9tvz" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.808368 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.817950 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.831603 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.851314 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.852765 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.859066 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-xt65f" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.859447 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860237 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csff5\" (UniqueName: \"kubernetes.io/projected/f245e2df-5488-4c94-a1ad-a3b687c7e2d4-kube-api-access-csff5\") pod \"telemetry-operator-controller-manager-5db5cf686f-mpf75\" (UID: \"f245e2df-5488-4c94-a1ad-a3b687c7e2d4\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860282 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860350 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwxv\" (UniqueName: \"kubernetes.io/projected/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-kube-api-access-mvwxv\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h42ml\" (UniqueName: \"kubernetes.io/projected/b4d484dc-fd46-432a-963d-049705db3803-kube-api-access-h42ml\") pod \"swift-operator-controller-manager-6859f9b676-rntmf\" (UID: \"b4d484dc-fd46-432a-963d-049705db3803\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjps4\" (UniqueName: \"kubernetes.io/projected/714d66a0-6e2e-4f57-8b39-39fff761851f-kube-api-access-mjps4\") pod \"placement-operator-controller-manager-7d8bb7f44c-rjzvl\" (UID: \"714d66a0-6e2e-4f57-8b39-39fff761851f\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.860419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlvg\" (UniqueName: \"kubernetes.io/projected/6aef6e41-8436-4b85-8bf8-befc6314fee0-kube-api-access-fwlvg\") pod \"ovn-operator-controller-manager-688db7b6c7-r8w7q\" (UID: \"6aef6e41-8436-4b85-8bf8-befc6314fee0\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:21 crc kubenswrapper[4959]: E1003 13:49:21.860675 4959 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:49:21 crc kubenswrapper[4959]: E1003 13:49:21.860707 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert podName:f8d0c10b-38a4-435e-a620-6e69d1a31d5c nodeName:}" failed. No retries permitted until 2025-10-03 13:49:22.360695626 +0000 UTC m=+1131.564039043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" (UID: "f8d0c10b-38a4-435e-a620-6e69d1a31d5c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.879526 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.897255 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.912713 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.922968 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.929157 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwxv\" (UniqueName: \"kubernetes.io/projected/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-kube-api-access-mvwxv\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.931184 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h42ml\" (UniqueName: \"kubernetes.io/projected/b4d484dc-fd46-432a-963d-049705db3803-kube-api-access-h42ml\") pod \"swift-operator-controller-manager-6859f9b676-rntmf\" (UID: \"b4d484dc-fd46-432a-963d-049705db3803\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.935410 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.940595 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-dhwxc" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.945756 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlvg\" (UniqueName: \"kubernetes.io/projected/6aef6e41-8436-4b85-8bf8-befc6314fee0-kube-api-access-fwlvg\") pod \"ovn-operator-controller-manager-688db7b6c7-r8w7q\" (UID: \"6aef6e41-8436-4b85-8bf8-befc6314fee0\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.963326 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjps4\" (UniqueName: \"kubernetes.io/projected/714d66a0-6e2e-4f57-8b39-39fff761851f-kube-api-access-mjps4\") pod \"placement-operator-controller-manager-7d8bb7f44c-rjzvl\" (UID: \"714d66a0-6e2e-4f57-8b39-39fff761851f\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.964358 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csff5\" (UniqueName: \"kubernetes.io/projected/f245e2df-5488-4c94-a1ad-a3b687c7e2d4-kube-api-access-csff5\") pod \"telemetry-operator-controller-manager-5db5cf686f-mpf75\" (UID: \"f245e2df-5488-4c94-a1ad-a3b687c7e2d4\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.964574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zhkb\" (UniqueName: \"kubernetes.io/projected/e0651c70-7f08-46e0-8ce8-580d5f41583e-kube-api-access-5zhkb\") pod \"test-operator-controller-manager-5cd5cb47d7-7kc7n\" (UID: \"e0651c70-7f08-46e0-8ce8-580d5f41583e\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.969289 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j"] Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.994245 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:21 crc kubenswrapper[4959]: I1003 13:49:21.999946 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csff5\" (UniqueName: \"kubernetes.io/projected/f245e2df-5488-4c94-a1ad-a3b687c7e2d4-kube-api-access-csff5\") pod \"telemetry-operator-controller-manager-5db5cf686f-mpf75\" (UID: \"f245e2df-5488-4c94-a1ad-a3b687c7e2d4\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.053256 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw"] Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.054347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw"] Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.054434 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.058558 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zlc88" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.058741 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.073267 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc"] Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.074021 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.074257 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zhkb\" (UniqueName: \"kubernetes.io/projected/e0651c70-7f08-46e0-8ce8-580d5f41583e-kube-api-access-5zhkb\") pod \"test-operator-controller-manager-5cd5cb47d7-7kc7n\" (UID: \"e0651c70-7f08-46e0-8ce8-580d5f41583e\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.074294 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2z5\" (UniqueName: \"kubernetes.io/projected/9e10f72a-6b02-42c9-aa42-7394e8afac6b-kube-api-access-5s2z5\") pod \"watcher-operator-controller-manager-fcd7d9895-lsw9j\" (UID: \"9e10f72a-6b02-42c9-aa42-7394e8afac6b\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.074340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.085619 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-smstk" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.106394 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc"] Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.106660 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.113066 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zhkb\" (UniqueName: \"kubernetes.io/projected/e0651c70-7f08-46e0-8ce8-580d5f41583e-kube-api-access-5zhkb\") pod \"test-operator-controller-manager-5cd5cb47d7-7kc7n\" (UID: \"e0651c70-7f08-46e0-8ce8-580d5f41583e\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.125744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e06190ac-1517-453b-a038-8437d0c77ee1-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-cg44n\" (UID: \"e06190ac-1517-453b-a038-8437d0c77ee1\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.178110 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2z5\" (UniqueName: \"kubernetes.io/projected/9e10f72a-6b02-42c9-aa42-7394e8afac6b-kube-api-access-5s2z5\") pod \"watcher-operator-controller-manager-fcd7d9895-lsw9j\" (UID: \"9e10f72a-6b02-42c9-aa42-7394e8afac6b\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.178166 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.178229 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgsdl\" (UniqueName: \"kubernetes.io/projected/6165adf7-aef8-42ec-9ae4-662ade35b9fd-kube-api-access-kgsdl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc\" (UID: \"6165adf7-aef8-42ec-9ae4-662ade35b9fd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.178287 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5s7s\" (UniqueName: \"kubernetes.io/projected/f3e29379-18ac-4401-be88-623abc26373e-kube-api-access-b5s7s\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.196163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2z5\" (UniqueName: \"kubernetes.io/projected/9e10f72a-6b02-42c9-aa42-7394e8afac6b-kube-api-access-5s2z5\") pod \"watcher-operator-controller-manager-fcd7d9895-lsw9j\" (UID: \"9e10f72a-6b02-42c9-aa42-7394e8afac6b\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.212055 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.218781 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.235995 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.249436 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.283477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.283530 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgsdl\" (UniqueName: \"kubernetes.io/projected/6165adf7-aef8-42ec-9ae4-662ade35b9fd-kube-api-access-kgsdl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc\" (UID: \"6165adf7-aef8-42ec-9ae4-662ade35b9fd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.283586 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5s7s\" (UniqueName: \"kubernetes.io/projected/f3e29379-18ac-4401-be88-623abc26373e-kube-api-access-b5s7s\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.284041 4959 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.284087 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert podName:f3e29379-18ac-4401-be88-623abc26373e nodeName:}" failed. No retries permitted until 2025-10-03 13:49:22.784072653 +0000 UTC m=+1131.987416070 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert") pod "openstack-operator-controller-manager-8fd589ff7-nf5sw" (UID: "f3e29379-18ac-4401-be88-623abc26373e") : secret "webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.315448 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.333606 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgsdl\" (UniqueName: \"kubernetes.io/projected/6165adf7-aef8-42ec-9ae4-662ade35b9fd-kube-api-access-kgsdl\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc\" (UID: \"6165adf7-aef8-42ec-9ae4-662ade35b9fd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.362146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5s7s\" (UniqueName: \"kubernetes.io/projected/f3e29379-18ac-4401-be88-623abc26373e-kube-api-access-b5s7s\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.387265 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.388091 4959 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.388148 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert podName:f8d0c10b-38a4-435e-a620-6e69d1a31d5c nodeName:}" failed. No retries permitted until 2025-10-03 13:49:23.388130311 +0000 UTC m=+1132.591473728 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" (UID: "f8d0c10b-38a4-435e-a620-6e69d1a31d5c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.395153 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6"] Oct 03 13:49:22 crc kubenswrapper[4959]: W1003 13:49:22.457728 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1848147c_d228_4d3c_bace_7628a688438d.slice/crio-f442c791b69a3bdf6e0083d6599830e128487110a9b432049ca406f38d92620d WatchSource:0}: Error finding container f442c791b69a3bdf6e0083d6599830e128487110a9b432049ca406f38d92620d: Status 404 returned error can't find the container with id f442c791b69a3bdf6e0083d6599830e128487110a9b432049ca406f38d92620d Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.463852 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.498379 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" event={"ID":"1848147c-d228-4d3c-bace-7628a688438d","Type":"ContainerStarted","Data":"f442c791b69a3bdf6e0083d6599830e128487110a9b432049ca406f38d92620d"} Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.654720 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.800503 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.802116 4959 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: E1003 13:49:22.802164 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert podName:f3e29379-18ac-4401-be88-623abc26373e nodeName:}" failed. No retries permitted until 2025-10-03 13:49:23.802151603 +0000 UTC m=+1133.005495020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert") pod "openstack-operator-controller-manager-8fd589ff7-nf5sw" (UID: "f3e29379-18ac-4401-be88-623abc26373e") : secret "webhook-server-cert" not found Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.849143 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs"] Oct 03 13:49:22 crc kubenswrapper[4959]: I1003 13:49:22.885952 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.060289 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.074020 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fdhll"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.241345 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.305214 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.314453 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod111756cf_9060_4c4b_878b_4ac998f89786.slice/crio-6511fbadbc935f51aa9274def0a13c351c8f49980ec772bfa73955d4f9fd6c15 WatchSource:0}: Error finding container 6511fbadbc935f51aa9274def0a13c351c8f49980ec772bfa73955d4f9fd6c15: Status 404 returned error can't find the container with id 6511fbadbc935f51aa9274def0a13c351c8f49980ec772bfa73955d4f9fd6c15 Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.324105 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.336263 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.338223 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aa2a2f3_7729_43a4_8564_c74ef4ddf86d.slice/crio-9ec73e7864d276881914bf906985cfa7920937b482023a5c238023ca867fbb38 WatchSource:0}: Error finding container 9ec73e7864d276881914bf906985cfa7920937b482023a5c238023ca867fbb38: Status 404 returned error can't find the container with id 9ec73e7864d276881914bf906985cfa7920937b482023a5c238023ca867fbb38 Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.339043 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod469d9704_d797_4ab2_9e51_b4a2a62942b7.slice/crio-99b96563519476b118d7884d494c42a5b4ba93e94aac2152ba895a823d72fea4 WatchSource:0}: Error finding container 99b96563519476b118d7884d494c42a5b4ba93e94aac2152ba895a823d72fea4: Status 404 returned error can't find the container with id 99b96563519476b118d7884d494c42a5b4ba93e94aac2152ba895a823d72fea4 Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.347275 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.350534 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.354666 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.361847 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f49a60f_299a_4568_81ca_687eb8d439d3.slice/crio-ae6019f815db4e20f0c90611b6672e7a5f1cdbc917b0c463421880c2d10b2b00 WatchSource:0}: Error finding container ae6019f815db4e20f0c90611b6672e7a5f1cdbc917b0c463421880c2d10b2b00: Status 404 returned error can't find the container with id ae6019f815db4e20f0c90611b6672e7a5f1cdbc917b0c463421880c2d10b2b00 Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.414117 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.427108 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f8d0c10b-38a4-435e-a620-6e69d1a31d5c-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678hr24c\" (UID: \"f8d0c10b-38a4-435e-a620-6e69d1a31d5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.515442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" event={"ID":"b991eb57-3139-462c-921b-72525e281d25","Type":"ContainerStarted","Data":"e28c25dee2f117c3670d4f87ad6a99c454789db9b221481151077b7de5c9d06f"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.519373 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" event={"ID":"c5519a95-6c42-4bc5-be67-626e8878421c","Type":"ContainerStarted","Data":"836862ed8dcb6a00d7053814750c93cc5059c2eca88f9d860cc3023169b0001f"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.539331 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" event={"ID":"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6","Type":"ContainerStarted","Data":"dfa735abd9f67b52a90c3f21e824577d7b789efa39ece9cf15778dd4196a7c98"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.546466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" event={"ID":"37c195cb-ee19-4040-ae6a-388685fed18d","Type":"ContainerStarted","Data":"cc78fc51bfa0430845d6af4b4b36f2e3e6eea306ad908d94c8c93dea68b1fad6"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.550126 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" event={"ID":"84bc4832-d055-4c00-955d-72e589902ea7","Type":"ContainerStarted","Data":"06f78a5cbb1f8588f41360c459df477c93a0f360bc4328ed96eab485df030565"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.550985 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.553977 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.554483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" event={"ID":"111756cf-9060-4c4b-878b-4ac998f89786","Type":"ContainerStarted","Data":"6511fbadbc935f51aa9274def0a13c351c8f49980ec772bfa73955d4f9fd6c15"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.557065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" event={"ID":"0f49a60f-299a-4568-81ca-687eb8d439d3","Type":"ContainerStarted","Data":"ae6019f815db4e20f0c90611b6672e7a5f1cdbc917b0c463421880c2d10b2b00"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.559518 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" event={"ID":"ce1839e2-3240-40dd-9145-867852c68de2","Type":"ContainerStarted","Data":"8863490e7b679762fac4800f9d6d9bf7aa1f3a640802c266b3336437a3c42c0b"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.561069 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" event={"ID":"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d","Type":"ContainerStarted","Data":"9ec73e7864d276881914bf906985cfa7920937b482023a5c238023ca867fbb38"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.562135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" event={"ID":"f1efc6bb-e4c4-437f-8a29-8b587327652b","Type":"ContainerStarted","Data":"0db0bd68a10ececba162d084315596f7b10acd04e45abfc3079a4bb5b20b778e"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.562863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" event={"ID":"469d9704-d797-4ab2-9e51-b4a2a62942b7","Type":"ContainerStarted","Data":"99b96563519476b118d7884d494c42a5b4ba93e94aac2152ba895a823d72fea4"} Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.716936 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.723679 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.728212 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.742337 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e10f72a_6b02_42c9_aa42_7394e8afac6b.slice/crio-d877c50995d9c12075b2fd0103b7b0700149ef576ff21cd118b31c7317a205b4 WatchSource:0}: Error finding container d877c50995d9c12075b2fd0103b7b0700149ef576ff21cd118b31c7317a205b4: Status 404 returned error can't find the container with id d877c50995d9c12075b2fd0103b7b0700149ef576ff21cd118b31c7317a205b4 Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.751632 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.753037 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod714d66a0_6e2e_4f57_8b39_39fff761851f.slice/crio-d5c8fbb4073e841d867fb2df476087c94a16e55cd8dc7e7dcbc7f384a9d9118c WatchSource:0}: Error finding container d5c8fbb4073e841d867fb2df476087c94a16e55cd8dc7e7dcbc7f384a9d9118c: Status 404 returned error can't find the container with id d5c8fbb4073e841d867fb2df476087c94a16e55cd8dc7e7dcbc7f384a9d9118c Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.756335 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.757618 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13a6dcff_d726_469a_9807_0ff2d76a7713.slice/crio-0447169a2a03067b0787a2a10d0ecc623d75eb355cb02d44624c10ac98a0b361 WatchSource:0}: Error finding container 0447169a2a03067b0787a2a10d0ecc623d75eb355cb02d44624c10ac98a0b361: Status 404 returned error can't find the container with id 0447169a2a03067b0787a2a10d0ecc623d75eb355cb02d44624c10ac98a0b361 Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.761425 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6165adf7_aef8_42ec_9ae4_662ade35b9fd.slice/crio-bd1ef361b2270bdeb3504e38253bc2ac75aea751f897acd24ce65b3451310d12 WatchSource:0}: Error finding container bd1ef361b2270bdeb3504e38253bc2ac75aea751f897acd24ce65b3451310d12: Status 404 returned error can't find the container with id bd1ef361b2270bdeb3504e38253bc2ac75aea751f897acd24ce65b3451310d12 Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.761679 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jrs2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5c468bf4d4-qkmbn_openstack-operators(13a6dcff-d726-469a-9807-0ff2d76a7713): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.763665 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgsdl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc_openstack-operators(6165adf7-aef8-42ec-9ae4-662ade35b9fd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.764757 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" podUID="6165adf7-aef8-42ec-9ae4-662ade35b9fd" Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.765173 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csff5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5db5cf686f-mpf75_openstack-operators(f245e2df-5488-4c94-a1ad-a3b687c7e2d4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.765933 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc"] Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.770907 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75"] Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.772145 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0651c70_7f08_46e0_8ce8_580d5f41583e.slice/crio-caebc9a90cf7964a2687bf1b6b96afd1908dcc6cba4bd8d516a8f048fbf3b866 WatchSource:0}: Error finding container caebc9a90cf7964a2687bf1b6b96afd1908dcc6cba4bd8d516a8f048fbf3b866: Status 404 returned error can't find the container with id caebc9a90cf7964a2687bf1b6b96afd1908dcc6cba4bd8d516a8f048fbf3b866 Oct 03 13:49:23 crc kubenswrapper[4959]: W1003 13:49:23.774101 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode06190ac_1517_453b_a038_8437d0c77ee1.slice/crio-1c007f6b762fa779aed9239a23774e3d63b959b9756a6711b7d41749bc8f540c WatchSource:0}: Error finding container 1c007f6b762fa779aed9239a23774e3d63b959b9756a6711b7d41749bc8f540c: Status 404 returned error can't find the container with id 1c007f6b762fa779aed9239a23774e3d63b959b9756a6711b7d41749bc8f540c Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.774128 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5zhkb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-7kc7n_openstack-operators(e0651c70-7f08-46e0-8ce8-580d5f41583e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.775114 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n"] Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.779412 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sclfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5fbf469cd7-cg44n_openstack-operators(e06190ac-1517-453b-a038-8437d0c77ee1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: E1003 13:49:23.781550 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h42ml,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-rntmf_openstack-operators(b4d484dc-fd46-432a-963d-049705db3803): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.819497 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:23 crc kubenswrapper[4959]: I1003 13:49:23.824115 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e29379-18ac-4401-be88-623abc26373e-cert\") pod \"openstack-operator-controller-manager-8fd589ff7-nf5sw\" (UID: \"f3e29379-18ac-4401-be88-623abc26373e\") " pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.029252 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c"] Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.072848 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.505028 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw"] Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.583911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" event={"ID":"e06190ac-1517-453b-a038-8437d0c77ee1","Type":"ContainerStarted","Data":"1c007f6b762fa779aed9239a23774e3d63b959b9756a6711b7d41749bc8f540c"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.586633 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" event={"ID":"f245e2df-5488-4c94-a1ad-a3b687c7e2d4","Type":"ContainerStarted","Data":"c6be3165cda4459203a5afd0b0f11a348edf97e2338f75c42a3b0f4ce3b74085"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.587916 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" event={"ID":"6165adf7-aef8-42ec-9ae4-662ade35b9fd","Type":"ContainerStarted","Data":"bd1ef361b2270bdeb3504e38253bc2ac75aea751f897acd24ce65b3451310d12"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.588784 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" event={"ID":"e0651c70-7f08-46e0-8ce8-580d5f41583e","Type":"ContainerStarted","Data":"caebc9a90cf7964a2687bf1b6b96afd1908dcc6cba4bd8d516a8f048fbf3b866"} Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.589523 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" podUID="6165adf7-aef8-42ec-9ae4-662ade35b9fd" Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.590161 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" event={"ID":"6aef6e41-8436-4b85-8bf8-befc6314fee0","Type":"ContainerStarted","Data":"2fcd50e5d360c29d4f250889fbcf4f9fe272c66c324c65fba8c76dadf09d7194"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.591554 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" event={"ID":"13a6dcff-d726-469a-9807-0ff2d76a7713","Type":"ContainerStarted","Data":"0447169a2a03067b0787a2a10d0ecc623d75eb355cb02d44624c10ac98a0b361"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.592955 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" event={"ID":"714d66a0-6e2e-4f57-8b39-39fff761851f","Type":"ContainerStarted","Data":"d5c8fbb4073e841d867fb2df476087c94a16e55cd8dc7e7dcbc7f384a9d9118c"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.594298 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" event={"ID":"b4d484dc-fd46-432a-963d-049705db3803","Type":"ContainerStarted","Data":"dcdbfeeb799e6b8d055eeb50e52fd7617fe5448bd0536d0b307dc04f0f30e859"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.595667 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" event={"ID":"f3e29379-18ac-4401-be88-623abc26373e","Type":"ContainerStarted","Data":"294c589c1767551b47330514018414f5fe3d775c25bce8898b1bb696fac13505"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.596766 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" event={"ID":"9e10f72a-6b02-42c9-aa42-7394e8afac6b","Type":"ContainerStarted","Data":"d877c50995d9c12075b2fd0103b7b0700149ef576ff21cd118b31c7317a205b4"} Oct 03 13:49:24 crc kubenswrapper[4959]: I1003 13:49:24.598058 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" event={"ID":"f8d0c10b-38a4-435e-a620-6e69d1a31d5c","Type":"ContainerStarted","Data":"a0c7f9648dcd7815831f1086b0d7e375004c27717f39805aa6682b70f5478e6e"} Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.761140 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" podUID="e0651c70-7f08-46e0-8ce8-580d5f41583e" Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.762744 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" podUID="f245e2df-5488-4c94-a1ad-a3b687c7e2d4" Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.766380 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" podUID="e06190ac-1517-453b-a038-8437d0c77ee1" Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.772348 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" podUID="b4d484dc-fd46-432a-963d-049705db3803" Oct 03 13:49:24 crc kubenswrapper[4959]: E1003 13:49:24.774753 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" podUID="13a6dcff-d726-469a-9807-0ff2d76a7713" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.611147 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" event={"ID":"f3e29379-18ac-4401-be88-623abc26373e","Type":"ContainerStarted","Data":"0eb960e10d7ab6aafe067c24d48e8d9460fffeec82785757876424114c86fb91"} Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.611488 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" event={"ID":"f3e29379-18ac-4401-be88-623abc26373e","Type":"ContainerStarted","Data":"28e2ecc1e5f923ea2d069b8fd63a8a316ccec495b03b77b2eb33e0e0c7b158a8"} Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.611587 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.620591 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" event={"ID":"f245e2df-5488-4c94-a1ad-a3b687c7e2d4","Type":"ContainerStarted","Data":"f2ae37a6e868c212b53b64602a9087a432bf1426a222ebec8d148d72ca1f34e2"} Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.623229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" event={"ID":"e0651c70-7f08-46e0-8ce8-580d5f41583e","Type":"ContainerStarted","Data":"ff060c35cf8ea4f3996119b2c6cd35875f33ddb31649e33f48733929e3c271dc"} Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.626158 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" podUID="e0651c70-7f08-46e0-8ce8-580d5f41583e" Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.626278 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" podUID="f245e2df-5488-4c94-a1ad-a3b687c7e2d4" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.627636 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" event={"ID":"13a6dcff-d726-469a-9807-0ff2d76a7713","Type":"ContainerStarted","Data":"fa9e99a82ddd447eb7fd9b7d2933fb50d694bb58258df557c67697d26f6ff757"} Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.628991 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" podUID="13a6dcff-d726-469a-9807-0ff2d76a7713" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.630334 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" event={"ID":"e06190ac-1517-453b-a038-8437d0c77ee1","Type":"ContainerStarted","Data":"59636adb218d9b8e5b2c166d85e78d629efb0e85225c7f0a0c61ac884e864a60"} Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.631770 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" podUID="e06190ac-1517-453b-a038-8437d0c77ee1" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.635211 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" event={"ID":"b4d484dc-fd46-432a-963d-049705db3803","Type":"ContainerStarted","Data":"8945cf6789a48a04b593cd68a60842e31895a2c6c05b950f0760e001d55b641c"} Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.640240 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" podUID="b4d484dc-fd46-432a-963d-049705db3803" Oct 03 13:49:25 crc kubenswrapper[4959]: E1003 13:49:25.640405 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" podUID="6165adf7-aef8-42ec-9ae4-662ade35b9fd" Oct 03 13:49:25 crc kubenswrapper[4959]: I1003 13:49:25.647307 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" podStartSLOduration=4.647285831 podStartE2EDuration="4.647285831s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:49:25.639730658 +0000 UTC m=+1134.843074075" watchObservedRunningTime="2025-10-03 13:49:25.647285831 +0000 UTC m=+1134.850629248" Oct 03 13:49:26 crc kubenswrapper[4959]: E1003 13:49:26.644915 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" podUID="e0651c70-7f08-46e0-8ce8-580d5f41583e" Oct 03 13:49:26 crc kubenswrapper[4959]: E1003 13:49:26.644973 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" podUID="b4d484dc-fd46-432a-963d-049705db3803" Oct 03 13:49:26 crc kubenswrapper[4959]: E1003 13:49:26.646009 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" podUID="f245e2df-5488-4c94-a1ad-a3b687c7e2d4" Oct 03 13:49:26 crc kubenswrapper[4959]: E1003 13:49:26.646052 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" podUID="e06190ac-1517-453b-a038-8437d0c77ee1" Oct 03 13:49:26 crc kubenswrapper[4959]: E1003 13:49:26.651068 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" podUID="13a6dcff-d726-469a-9807-0ff2d76a7713" Oct 03 13:49:34 crc kubenswrapper[4959]: I1003 13:49:34.077574 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8fd589ff7-nf5sw" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.044663 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.044981 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.045028 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.045652 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.045710 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694" gracePeriod=600 Oct 03 13:49:36 crc kubenswrapper[4959]: E1003 13:49:36.174928 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5" Oct 03 13:49:36 crc kubenswrapper[4959]: E1003 13:49:36.175086 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6rkq5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-599898f689-fdhll_openstack-operators(84bc4832-d055-4c00-955d-72e589902ea7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:49:36 crc kubenswrapper[4959]: E1003 13:49:36.547870 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" podUID="84bc4832-d055-4c00-955d-72e589902ea7" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.723017 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694" exitCode=0 Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.723094 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694"} Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.723128 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77"} Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.723149 4959 scope.go:117] "RemoveContainer" containerID="abedbf766594f4a4b5cb1d0a3a6c0a6d80d526e895ffe7358c0a81d667bb1c2a" Oct 03 13:49:36 crc kubenswrapper[4959]: I1003 13:49:36.731035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" event={"ID":"84bc4832-d055-4c00-955d-72e589902ea7","Type":"ContainerStarted","Data":"67815ba1aa53b0c06cfe1b77b742a699dadfbd42b39ccce4e41e465dd0a73388"} Oct 03 13:49:36 crc kubenswrapper[4959]: E1003 13:49:36.732525 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5\\\"\"" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" podUID="84bc4832-d055-4c00-955d-72e589902ea7" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.754407 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" event={"ID":"ce1839e2-3240-40dd-9145-867852c68de2","Type":"ContainerStarted","Data":"12b7759844d6dbf9ec81724d6cca502443f52b676e72a9fae577fadfe5bbc811"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.754947 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" event={"ID":"ce1839e2-3240-40dd-9145-867852c68de2","Type":"ContainerStarted","Data":"80061cc20695e29b7dacabf97e3ab781785a2afcd60aa3836a31ab800ecb547c"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.756000 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.758760 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" event={"ID":"6aef6e41-8436-4b85-8bf8-befc6314fee0","Type":"ContainerStarted","Data":"0d31e84f833a85c953ab9d835c1c95ec6e4ab976e9a15db288e21f2a2990f695"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.758798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" event={"ID":"6aef6e41-8436-4b85-8bf8-befc6314fee0","Type":"ContainerStarted","Data":"9118dbb19e57bd80069fc6955b453759d3e749d2a848a62471797b77cd4bdba8"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.759410 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.765843 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" event={"ID":"1848147c-d228-4d3c-bace-7628a688438d","Type":"ContainerStarted","Data":"7eb5db4051c6308b52f069fe4ae4bfdc6b1aaea7c02172dbd4b872675378c316"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.775143 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" event={"ID":"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d","Type":"ContainerStarted","Data":"dd250b21b4ab0e58a65837e3d5dc5fe3946996fbb7cd002e53599b9e4672b5b8"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.775941 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.784592 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" podStartSLOduration=3.860987374 podStartE2EDuration="16.784570387s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.334354102 +0000 UTC m=+1132.537697519" lastFinishedPulling="2025-10-03 13:49:36.257937105 +0000 UTC m=+1145.461280532" observedRunningTime="2025-10-03 13:49:37.777432075 +0000 UTC m=+1146.980775492" watchObservedRunningTime="2025-10-03 13:49:37.784570387 +0000 UTC m=+1146.987913804" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.797527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" event={"ID":"37c195cb-ee19-4040-ae6a-388685fed18d","Type":"ContainerStarted","Data":"15765177934ec027d5e03697b474d8f0d85c0a2f78126ccb08d1ae6f95cd717b"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.808000 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" event={"ID":"0f49a60f-299a-4568-81ca-687eb8d439d3","Type":"ContainerStarted","Data":"9f7b95375bb34079ff893f5059b5ded775c46d836d063b04df83cd63b9e1b63a"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.808052 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" event={"ID":"0f49a60f-299a-4568-81ca-687eb8d439d3","Type":"ContainerStarted","Data":"02baec380df99bcbd09ae2b6a8e971f11f253f1f8748f8fb619f02a3bdd90bad"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.808656 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.817009 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" podStartSLOduration=3.8728860210000002 podStartE2EDuration="16.816985739s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.341395922 +0000 UTC m=+1132.544739339" lastFinishedPulling="2025-10-03 13:49:36.28549563 +0000 UTC m=+1145.488839057" observedRunningTime="2025-10-03 13:49:37.811443425 +0000 UTC m=+1147.014786862" watchObservedRunningTime="2025-10-03 13:49:37.816985739 +0000 UTC m=+1147.020329156" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.845902 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" event={"ID":"b991eb57-3139-462c-921b-72525e281d25","Type":"ContainerStarted","Data":"ee1226de74c9e03a03b71da584f6635122a503e1f5113167045705c96829ffbe"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.869659 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" podStartSLOduration=3.95576572 podStartE2EDuration="16.869642378s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.363776142 +0000 UTC m=+1132.567119559" lastFinishedPulling="2025-10-03 13:49:36.2776528 +0000 UTC m=+1145.480996217" observedRunningTime="2025-10-03 13:49:37.867515907 +0000 UTC m=+1147.070859314" watchObservedRunningTime="2025-10-03 13:49:37.869642378 +0000 UTC m=+1147.072985795" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.872323 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" podStartSLOduration=4.144867149 podStartE2EDuration="16.872312183s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.554479369 +0000 UTC m=+1132.757822806" lastFinishedPulling="2025-10-03 13:49:36.281924423 +0000 UTC m=+1145.485267840" observedRunningTime="2025-10-03 13:49:37.841125151 +0000 UTC m=+1147.044468568" watchObservedRunningTime="2025-10-03 13:49:37.872312183 +0000 UTC m=+1147.075655590" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.890844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" event={"ID":"c5519a95-6c42-4bc5-be67-626e8878421c","Type":"ContainerStarted","Data":"eed2ca2b967a49ade28fb668ca5aa752b3645efdd6dda79a0c399e82cae4709e"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.891951 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.901864 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" event={"ID":"f8d0c10b-38a4-435e-a620-6e69d1a31d5c","Type":"ContainerStarted","Data":"37ecf3960852368ab1f43f1b8e75c85b25374a6fc6b8e1b3b4cabc12df0de64a"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.915147 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" event={"ID":"714d66a0-6e2e-4f57-8b39-39fff761851f","Type":"ContainerStarted","Data":"39d05676d832d0839a46542a1fba65440dade6fff6c334af06406ea245b012ef"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.926085 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" podStartSLOduration=3.549548736 podStartE2EDuration="16.926061378s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:22.906084168 +0000 UTC m=+1132.109427585" lastFinishedPulling="2025-10-03 13:49:36.28259681 +0000 UTC m=+1145.485940227" observedRunningTime="2025-10-03 13:49:37.920272949 +0000 UTC m=+1147.123616366" watchObservedRunningTime="2025-10-03 13:49:37.926061378 +0000 UTC m=+1147.129404795" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.926383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" event={"ID":"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6","Type":"ContainerStarted","Data":"946b99714ec7d64d0be14a69a268652a43389d7e4b00186f44bbd2a999e86e03"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.927273 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.939425 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" event={"ID":"f1efc6bb-e4c4-437f-8a29-8b587327652b","Type":"ContainerStarted","Data":"b22ce5161873c77579b81e1cffec51148541095f1ca25c7aa1d6ba1137b8486a"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.940026 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.956883 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" event={"ID":"469d9704-d797-4ab2-9e51-b4a2a62942b7","Type":"ContainerStarted","Data":"02644671ab6afc2654a06e5cbc2321b5b9d4c16f9212c979c91f40419b5e2c6e"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.956924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" event={"ID":"469d9704-d797-4ab2-9e51-b4a2a62942b7","Type":"ContainerStarted","Data":"026eb2852a2a63454caf95894124fc2b6ea7ea1fd0000540fd2050633f74065f"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.957589 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.978604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" event={"ID":"111756cf-9060-4c4b-878b-4ac998f89786","Type":"ContainerStarted","Data":"5ff8ca25e34361d6c55ada0e8687b4976b89e7c49816ef29975335a451a13a7e"} Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.979649 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.987826 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" podStartSLOduration=3.76827488 podStartE2EDuration="16.987809427s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.062786075 +0000 UTC m=+1132.266129492" lastFinishedPulling="2025-10-03 13:49:36.282320612 +0000 UTC m=+1145.485664039" observedRunningTime="2025-10-03 13:49:37.952564877 +0000 UTC m=+1147.155908294" watchObservedRunningTime="2025-10-03 13:49:37.987809427 +0000 UTC m=+1147.191152844" Oct 03 13:49:37 crc kubenswrapper[4959]: I1003 13:49:37.989142 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" podStartSLOduration=3.600662409 podStartE2EDuration="16.989135029s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:22.893580676 +0000 UTC m=+1132.096924093" lastFinishedPulling="2025-10-03 13:49:36.282053286 +0000 UTC m=+1145.485396713" observedRunningTime="2025-10-03 13:49:37.986622939 +0000 UTC m=+1147.189966376" watchObservedRunningTime="2025-10-03 13:49:37.989135029 +0000 UTC m=+1147.192478446" Oct 03 13:49:38 crc kubenswrapper[4959]: I1003 13:49:38.006670 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" event={"ID":"9e10f72a-6b02-42c9-aa42-7394e8afac6b","Type":"ContainerStarted","Data":"e563f1f58ae9a77cfc79dfc38b0e522d7076dc59d840833ea4a2cf21c4f3e18d"} Oct 03 13:49:38 crc kubenswrapper[4959]: I1003 13:49:38.007017 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:38 crc kubenswrapper[4959]: E1003 13:49:38.014279 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5\\\"\"" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" podUID="84bc4832-d055-4c00-955d-72e589902ea7" Oct 03 13:49:38 crc kubenswrapper[4959]: I1003 13:49:38.042611 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" podStartSLOduration=4.077336122 podStartE2EDuration="17.042595719s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.316716397 +0000 UTC m=+1132.520059824" lastFinishedPulling="2025-10-03 13:49:36.281975994 +0000 UTC m=+1145.485319421" observedRunningTime="2025-10-03 13:49:38.020520656 +0000 UTC m=+1147.223864083" watchObservedRunningTime="2025-10-03 13:49:38.042595719 +0000 UTC m=+1147.245939136" Oct 03 13:49:38 crc kubenswrapper[4959]: I1003 13:49:38.043603 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" podStartSLOduration=4.122355106 podStartE2EDuration="17.043598572s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.342781155 +0000 UTC m=+1132.546124572" lastFinishedPulling="2025-10-03 13:49:36.264024611 +0000 UTC m=+1145.467368038" observedRunningTime="2025-10-03 13:49:38.03687929 +0000 UTC m=+1147.240222707" watchObservedRunningTime="2025-10-03 13:49:38.043598572 +0000 UTC m=+1147.246941989" Oct 03 13:49:38 crc kubenswrapper[4959]: I1003 13:49:38.087173 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" podStartSLOduration=4.5459533180000005 podStartE2EDuration="17.087152782s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.745469473 +0000 UTC m=+1132.948812890" lastFinishedPulling="2025-10-03 13:49:36.286668927 +0000 UTC m=+1145.490012354" observedRunningTime="2025-10-03 13:49:38.085742598 +0000 UTC m=+1147.289086015" watchObservedRunningTime="2025-10-03 13:49:38.087152782 +0000 UTC m=+1147.290496199" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.028255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" event={"ID":"9e10f72a-6b02-42c9-aa42-7394e8afac6b","Type":"ContainerStarted","Data":"f8c435f094f21ad9e46ef270f40dc45169212250fd30c10ef156e0293ac1882a"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.031087 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" event={"ID":"6aa2a2f3-7729-43a4-8564-c74ef4ddf86d","Type":"ContainerStarted","Data":"16a03e999ff146574af826557850b5d04c60e2a6b8cb86a0caf4d9fa46d5a97f"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.033170 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" event={"ID":"714d66a0-6e2e-4f57-8b39-39fff761851f","Type":"ContainerStarted","Data":"6a180407b64d1d9c18ffe3bcb5cc54bae7dd5dc92b025ffbe4fba37d9c6ba372"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.033238 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.034573 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" event={"ID":"a1ef21c1-ce09-47cc-b1c6-a7e8916431c6","Type":"ContainerStarted","Data":"1fe1f7ae60a44691df4e7ab33e50b3a9a30f60fb42c2b3cc1eef0b8d40b4d611"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.040552 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" event={"ID":"b991eb57-3139-462c-921b-72525e281d25","Type":"ContainerStarted","Data":"6aefef6b51968d5ea8e200b1182e7b5178b61139f109c154408485088fa78025"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.040675 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.042799 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" event={"ID":"c5519a95-6c42-4bc5-be67-626e8878421c","Type":"ContainerStarted","Data":"51547b6ba39bdda144cd5b3e9f449be16d03bf719aa626193d1957155a01e99a"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.046131 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" event={"ID":"f8d0c10b-38a4-435e-a620-6e69d1a31d5c","Type":"ContainerStarted","Data":"bbee55ea31151071e6b48c46bdb7aedfdd2de030af685e7d3de70e1fa18e8e73"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.046244 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.054013 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" podStartSLOduration=5.529053909 podStartE2EDuration="18.053990351s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.757441952 +0000 UTC m=+1132.960785369" lastFinishedPulling="2025-10-03 13:49:36.282378374 +0000 UTC m=+1145.485721811" observedRunningTime="2025-10-03 13:49:39.050535777 +0000 UTC m=+1148.253879184" watchObservedRunningTime="2025-10-03 13:49:39.053990351 +0000 UTC m=+1148.257333768" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.054427 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" event={"ID":"111756cf-9060-4c4b-878b-4ac998f89786","Type":"ContainerStarted","Data":"7b79d9509059d8271de27f1d993932e91d9fd8bbae869bebcded6eab917deea7"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.059064 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" event={"ID":"1848147c-d228-4d3c-bace-7628a688438d","Type":"ContainerStarted","Data":"c6f051ba654e83ef36d892f903a89f65b1093364b6a27f3f4d8f8abb13f368c7"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.059218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.061722 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" event={"ID":"37c195cb-ee19-4040-ae6a-388685fed18d","Type":"ContainerStarted","Data":"05b47d005b6524a6caa09e66c130bfeda7783bc20ccedc84c19bbcea423412e6"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.061848 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.066534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" event={"ID":"f1efc6bb-e4c4-437f-8a29-8b587327652b","Type":"ContainerStarted","Data":"81ca0ab0673401f1045582013c8513e9b68547ff90bb415ff0d63816f8d6ca4d"} Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.072464 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" podStartSLOduration=5.044422304 podStartE2EDuration="18.072444005s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.256146927 +0000 UTC m=+1132.459490344" lastFinishedPulling="2025-10-03 13:49:36.284168607 +0000 UTC m=+1145.487512045" observedRunningTime="2025-10-03 13:49:39.066421981 +0000 UTC m=+1148.269765398" watchObservedRunningTime="2025-10-03 13:49:39.072444005 +0000 UTC m=+1148.275787432" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.103794 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" podStartSLOduration=5.86474022 podStartE2EDuration="18.10377648s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:24.0427432 +0000 UTC m=+1133.246086617" lastFinishedPulling="2025-10-03 13:49:36.28177945 +0000 UTC m=+1145.485122877" observedRunningTime="2025-10-03 13:49:39.09504056 +0000 UTC m=+1148.298383987" watchObservedRunningTime="2025-10-03 13:49:39.10377648 +0000 UTC m=+1148.307119897" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.117408 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" podStartSLOduration=4.295764968 podStartE2EDuration="18.117387889s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:22.463655572 +0000 UTC m=+1131.666998989" lastFinishedPulling="2025-10-03 13:49:36.285278473 +0000 UTC m=+1145.488621910" observedRunningTime="2025-10-03 13:49:39.114069429 +0000 UTC m=+1148.317412886" watchObservedRunningTime="2025-10-03 13:49:39.117387889 +0000 UTC m=+1148.320731306" Oct 03 13:49:39 crc kubenswrapper[4959]: I1003 13:49:39.134067 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" podStartSLOduration=5.18194041 podStartE2EDuration="18.13404321s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.333339528 +0000 UTC m=+1132.536682945" lastFinishedPulling="2025-10-03 13:49:36.285442318 +0000 UTC m=+1145.488785745" observedRunningTime="2025-10-03 13:49:39.128614229 +0000 UTC m=+1148.331957656" watchObservedRunningTime="2025-10-03 13:49:39.13404321 +0000 UTC m=+1148.337386627" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.087665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" event={"ID":"b4d484dc-fd46-432a-963d-049705db3803","Type":"ContainerStarted","Data":"7f0a0020714e05f3bb0dac1d3656a70286689b0f00f1590ae55000adbbb68646"} Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.089148 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.092741 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" event={"ID":"e0651c70-7f08-46e0-8ce8-580d5f41583e","Type":"ContainerStarted","Data":"be403446a77a46ff15f21fc91024fe5357ad085ff41652aaef63262a9cb88002"} Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.093290 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.105449 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" podStartSLOduration=3.918492932 podStartE2EDuration="20.105429576s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.78142781 +0000 UTC m=+1132.984771227" lastFinishedPulling="2025-10-03 13:49:39.968364454 +0000 UTC m=+1149.171707871" observedRunningTime="2025-10-03 13:49:41.103322245 +0000 UTC m=+1150.306665692" watchObservedRunningTime="2025-10-03 13:49:41.105429576 +0000 UTC m=+1150.308773003" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.125121 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" podStartSLOduration=3.93707523 podStartE2EDuration="20.12510254s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.773996391 +0000 UTC m=+1132.977339818" lastFinishedPulling="2025-10-03 13:49:39.962023711 +0000 UTC m=+1149.165367128" observedRunningTime="2025-10-03 13:49:41.12263275 +0000 UTC m=+1150.325976167" watchObservedRunningTime="2025-10-03 13:49:41.12510254 +0000 UTC m=+1150.328445957" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.442632 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-n78z6" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.472853 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-5bgvs" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.520330 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-zpl6z" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.655812 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-crg46" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.724241 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-9ptc9" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.741086 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-55vd5" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.864075 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-h29xd" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.883600 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-ssfjs" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.937848 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-rfvdf" Oct 03 13:49:41 crc kubenswrapper[4959]: I1003 13:49:41.997836 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-r8w7q" Oct 03 13:49:42 crc kubenswrapper[4959]: I1003 13:49:42.215636 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-rjzvl" Oct 03 13:49:42 crc kubenswrapper[4959]: I1003 13:49:42.320565 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-lsw9j" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.118821 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" event={"ID":"6165adf7-aef8-42ec-9ae4-662ade35b9fd","Type":"ContainerStarted","Data":"0cdb9bb3cf92558a8502a93793784dedc4fdb0bff3392c8fd955aa6c6507368d"} Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.125030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" event={"ID":"13a6dcff-d726-469a-9807-0ff2d76a7713","Type":"ContainerStarted","Data":"6b10b7cea19b46dc36507f93b0cab2b553f3a19a2a9cefdb9f68c899daa3420c"} Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.125251 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.128956 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" event={"ID":"e06190ac-1517-453b-a038-8437d0c77ee1","Type":"ContainerStarted","Data":"c0ebb544a8abfd179697fb053a7e304f5bcedd50cfc795e3efc6c38bfb1ce5fc"} Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.129167 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.139634 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc" podStartSLOduration=2.301013255 podStartE2EDuration="21.139614594s" podCreationTimestamp="2025-10-03 13:49:22 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.763513538 +0000 UTC m=+1132.966856955" lastFinishedPulling="2025-10-03 13:49:42.602114877 +0000 UTC m=+1151.805458294" observedRunningTime="2025-10-03 13:49:43.134889831 +0000 UTC m=+1152.338233258" watchObservedRunningTime="2025-10-03 13:49:43.139614594 +0000 UTC m=+1152.342958011" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.157044 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" podStartSLOduration=3.318028906 podStartE2EDuration="22.157024644s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.779313719 +0000 UTC m=+1132.982657136" lastFinishedPulling="2025-10-03 13:49:42.618309457 +0000 UTC m=+1151.821652874" observedRunningTime="2025-10-03 13:49:43.154901513 +0000 UTC m=+1152.358244950" watchObservedRunningTime="2025-10-03 13:49:43.157024644 +0000 UTC m=+1152.360368061" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.175335 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" podStartSLOduration=3.328858567 podStartE2EDuration="22.175318505s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.76152212 +0000 UTC m=+1132.964865537" lastFinishedPulling="2025-10-03 13:49:42.607982058 +0000 UTC m=+1151.811325475" observedRunningTime="2025-10-03 13:49:43.174541486 +0000 UTC m=+1152.377884903" watchObservedRunningTime="2025-10-03 13:49:43.175318505 +0000 UTC m=+1152.378661922" Oct 03 13:49:43 crc kubenswrapper[4959]: I1003 13:49:43.559540 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678hr24c" Oct 03 13:49:44 crc kubenswrapper[4959]: I1003 13:49:44.140858 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" event={"ID":"f245e2df-5488-4c94-a1ad-a3b687c7e2d4","Type":"ContainerStarted","Data":"268f0b3c72ff195533285b2c91792795282de5ae3d68d55fe9a9670963cff819"} Oct 03 13:49:44 crc kubenswrapper[4959]: I1003 13:49:44.141232 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:44 crc kubenswrapper[4959]: I1003 13:49:44.172928 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" podStartSLOduration=3.05688065 podStartE2EDuration="23.172912894s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.764723387 +0000 UTC m=+1132.968066804" lastFinishedPulling="2025-10-03 13:49:43.880755631 +0000 UTC m=+1153.084099048" observedRunningTime="2025-10-03 13:49:44.170266221 +0000 UTC m=+1153.373609658" watchObservedRunningTime="2025-10-03 13:49:44.172912894 +0000 UTC m=+1153.376256311" Oct 03 13:49:50 crc kubenswrapper[4959]: I1003 13:49:50.202433 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" event={"ID":"84bc4832-d055-4c00-955d-72e589902ea7","Type":"ContainerStarted","Data":"95834699152b530732fc474a35fc3e9f267d783cbc1d2385e2ad2257be856e53"} Oct 03 13:49:50 crc kubenswrapper[4959]: I1003 13:49:50.203578 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:49:50 crc kubenswrapper[4959]: I1003 13:49:50.226709 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" podStartSLOduration=3.138202751 podStartE2EDuration="29.226685706s" podCreationTimestamp="2025-10-03 13:49:21 +0000 UTC" firstStartedPulling="2025-10-03 13:49:23.066446683 +0000 UTC m=+1132.269790100" lastFinishedPulling="2025-10-03 13:49:49.154929638 +0000 UTC m=+1158.358273055" observedRunningTime="2025-10-03 13:49:50.219538642 +0000 UTC m=+1159.422882069" watchObservedRunningTime="2025-10-03 13:49:50.226685706 +0000 UTC m=+1159.430029123" Oct 03 13:49:51 crc kubenswrapper[4959]: I1003 13:49:51.493230 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-dsq2k" Oct 03 13:49:51 crc kubenswrapper[4959]: I1003 13:49:51.733321 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-w96jd" Oct 03 13:49:51 crc kubenswrapper[4959]: I1003 13:49:51.834711 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-qkmbn" Oct 03 13:49:52 crc kubenswrapper[4959]: I1003 13:49:52.110821 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-rntmf" Oct 03 13:49:52 crc kubenswrapper[4959]: I1003 13:49:52.221796 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-mpf75" Oct 03 13:49:52 crc kubenswrapper[4959]: I1003 13:49:52.251770 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-7kc7n" Oct 03 13:49:52 crc kubenswrapper[4959]: I1003 13:49:52.252480 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-cg44n" Oct 03 13:50:01 crc kubenswrapper[4959]: I1003 13:50:01.591118 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-599898f689-fdhll" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.592722 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.594535 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.598690 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.598982 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-s992j" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.600110 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.600208 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.610093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.641632 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.642877 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.644599 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.649634 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.709352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.709430 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.709467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98h7k\" (UniqueName: \"kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.709517 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.709544 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr5nw\" (UniqueName: \"kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.810934 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.810984 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr5nw\" (UniqueName: \"kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.811054 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.811086 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.811115 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98h7k\" (UniqueName: \"kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.811950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.812258 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.812539 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.830403 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr5nw\" (UniqueName: \"kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw\") pod \"dnsmasq-dns-78dd6ddcc-gx9x4\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.832324 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98h7k\" (UniqueName: \"kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k\") pod \"dnsmasq-dns-675f4bcbfc-wtv4p\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.912802 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:15 crc kubenswrapper[4959]: I1003 13:50:15.965279 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:16 crc kubenswrapper[4959]: I1003 13:50:16.250728 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:16 crc kubenswrapper[4959]: I1003 13:50:16.401319 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:16 crc kubenswrapper[4959]: I1003 13:50:16.403699 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" event={"ID":"22be8fae-cbe5-4573-af02-07525ab3444b","Type":"ContainerStarted","Data":"f337cbfdbd43d839aebfec1fa2a5255461bbd7342fcda37d0b286ea8d169a6b1"} Oct 03 13:50:17 crc kubenswrapper[4959]: I1003 13:50:17.413287 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" event={"ID":"ab295ee6-3a3d-449a-9bc5-9176cad79051","Type":"ContainerStarted","Data":"a350d82a55f8472b3fba347e8154050c8306e9121229ea02940531a5df424503"} Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.221895 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.247282 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.248441 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.291505 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.368854 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.368924 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d4bw\" (UniqueName: \"kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.368952 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.470201 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d4bw\" (UniqueName: \"kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.470246 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.470320 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.471241 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.471274 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.506464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d4bw\" (UniqueName: \"kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw\") pod \"dnsmasq-dns-666b6646f7-cr4bb\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.586691 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.666807 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.715272 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.716970 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.734054 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.884979 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.885389 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ssw\" (UniqueName: \"kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:18 crc kubenswrapper[4959]: I1003 13:50:18.885441 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.008313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ssw\" (UniqueName: \"kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.008381 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.008406 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.009263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.009264 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.052773 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ssw\" (UniqueName: \"kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw\") pod \"dnsmasq-dns-57d769cc4f-4xmqn\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.095810 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:19 crc kubenswrapper[4959]: W1003 13:50:19.112233 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2613de27_74fb_444d_b431_7c083c7fc5b3.slice/crio-020d66a7052bcb9a66ad5464c614925caecc426a4998c0114f9771ac0f4682d5 WatchSource:0}: Error finding container 020d66a7052bcb9a66ad5464c614925caecc426a4998c0114f9771ac0f4682d5: Status 404 returned error can't find the container with id 020d66a7052bcb9a66ad5464c614925caecc426a4998c0114f9771ac0f4682d5 Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.339368 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.457660 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" event={"ID":"2613de27-74fb-444d-b431-7c083c7fc5b3","Type":"ContainerStarted","Data":"020d66a7052bcb9a66ad5464c614925caecc426a4998c0114f9771ac0f4682d5"} Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.485226 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.486413 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.490640 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.490934 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.490954 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.491130 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.491804 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.491954 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.492632 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jzgmg" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.499716 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615671 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615867 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615911 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615925 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.615948 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwxb8\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.616278 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.616333 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.616357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.616383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.717981 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718069 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718132 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwxb8\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718227 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718256 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718278 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718302 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.718363 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.719092 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.719287 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.719394 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.719916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.721178 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.722080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.727702 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.728234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.749658 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.753213 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.763145 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwxb8\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.771491 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.817652 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.839744 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.841385 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.845273 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.845445 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.845670 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.845865 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.846624 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.846740 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-slqtw" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.854232 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.855752 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.861334 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921028 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921066 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921094 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgbm9\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921111 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921131 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921220 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921281 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:19 crc kubenswrapper[4959]: I1003 13:50:19.921310 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022816 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022929 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022946 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022962 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.022994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.023014 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgbm9\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.023032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.023049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.023351 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.028159 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.031563 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.032250 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.033600 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.035864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.042010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.042227 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.051750 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.052953 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.060351 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgbm9\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.068457 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.199943 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.321853 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:50:20 crc kubenswrapper[4959]: I1003 13:50:20.468512 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" event={"ID":"cb452d4d-d280-4035-9a47-d240d9051982","Type":"ContainerStarted","Data":"29196dbd48b0a5216868e904688cae97384694480636df893c60588c6ebb11b1"} Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.032279 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.038985 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.042388 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.054134 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.054973 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-csztw" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.055865 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.057553 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.057714 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.087321 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160306 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160402 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160429 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160449 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjbz\" (UniqueName: \"kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160468 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160519 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.160548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262099 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262176 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262219 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjbz\" (UniqueName: \"kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262248 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262321 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262362 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262412 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.262440 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.263099 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.263825 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.264090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.264628 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.264734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.270639 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.271321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.274163 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.285934 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.290319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjbz\" (UniqueName: \"kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz\") pod \"openstack-galera-0\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " pod="openstack/openstack-galera-0" Oct 03 13:50:21 crc kubenswrapper[4959]: I1003 13:50:21.387541 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.445243 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.446848 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.454055 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.454132 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.454306 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.454413 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-nq9zj" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.458704 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589390 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589573 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589633 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhljg\" (UniqueName: \"kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589847 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.589879 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.590004 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.590035 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.691907 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.691960 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692064 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692088 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhljg\" (UniqueName: \"kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692115 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692145 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692224 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.692251 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.693043 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.693145 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.693488 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.693780 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.694422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.696529 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.701771 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.703264 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.715455 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhljg\" (UniqueName: \"kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.732961 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.733664 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.733988 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.741374 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.741610 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.741745 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-zgm5s" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.744995 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.767894 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.797110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsh8j\" (UniqueName: \"kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.797166 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.797276 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.797347 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.797397 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.899077 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.899881 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsh8j\" (UniqueName: \"kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.899916 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.899952 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.900038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.906629 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.907154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.907734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.909526 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:22 crc kubenswrapper[4959]: I1003 13:50:22.918346 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsh8j\" (UniqueName: \"kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j\") pod \"memcached-0\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " pod="openstack/memcached-0" Oct 03 13:50:23 crc kubenswrapper[4959]: I1003 13:50:23.098304 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.495469 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerStarted","Data":"24c85da72d38b0d376a5a5ae6ab49af74c8e798db542889a1745ed029c262fb2"} Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.500684 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.504218 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.508758 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-d5nmb" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.547324 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.653664 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6xtw\" (UniqueName: \"kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw\") pod \"kube-state-metrics-0\" (UID: \"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a\") " pod="openstack/kube-state-metrics-0" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.755392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6xtw\" (UniqueName: \"kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw\") pod \"kube-state-metrics-0\" (UID: \"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a\") " pod="openstack/kube-state-metrics-0" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.780117 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6xtw\" (UniqueName: \"kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw\") pod \"kube-state-metrics-0\" (UID: \"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a\") " pod="openstack/kube-state-metrics-0" Oct 03 13:50:24 crc kubenswrapper[4959]: I1003 13:50:24.843058 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:50:27 crc kubenswrapper[4959]: I1003 13:50:27.383587 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:27.999435 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.000816 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.003362 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vb566" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.003765 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.003974 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.007089 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.040764 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.042855 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.047101 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106577 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106804 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106897 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106934 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlw9\" (UniqueName: \"kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.106993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208334 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlw9\" (UniqueName: \"kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208445 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txg4\" (UniqueName: \"kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208606 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208641 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208715 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208738 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208761 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.208782 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.209458 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.209494 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.209551 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.212016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.217046 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.227915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.230973 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlw9\" (UniqueName: \"kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9\") pod \"ovn-controller-k4vvh\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.309757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.309845 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.309864 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.309886 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.310091 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.310145 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txg4\" (UniqueName: \"kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.310143 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.310140 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.310183 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.311127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.314411 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.325441 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txg4\" (UniqueName: \"kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4\") pod \"ovn-controller-ovs-2bhp5\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.329354 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.377178 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.882680 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.883981 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.886709 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.886927 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.887281 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.887823 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-twbwh" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.888362 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 03 13:50:28 crc kubenswrapper[4959]: I1003 13:50:28.891657 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019501 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2j7v\" (UniqueName: \"kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019570 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019604 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019621 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019650 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019705 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.019733 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121207 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2j7v\" (UniqueName: \"kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121487 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121510 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121525 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121609 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.121635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.123848 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.124090 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.128492 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.128877 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.132806 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.132900 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.139097 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.141240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2j7v\" (UniqueName: \"kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.149329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:29 crc kubenswrapper[4959]: I1003 13:50:29.254749 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.304847 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.307682 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.309126 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.309702 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.309811 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8w8cb" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.310231 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.311759 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 03 13:50:32 crc kubenswrapper[4959]: W1003 13:50:32.362960 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c9b61e3_0ba5_4abf_ac37_37e5395e2d6d.slice/crio-8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce WatchSource:0}: Error finding container 8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce: Status 404 returned error can't find the container with id 8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388381 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zmsd\" (UniqueName: \"kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388451 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388528 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388565 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.388665 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.391591 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.391761 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98h7k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-wtv4p_openstack(ab295ee6-3a3d-449a-9bc5-9176cad79051): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.393337 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" podUID="ab295ee6-3a3d-449a-9bc5-9176cad79051" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.460583 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.460758 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sr5nw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-gx9x4_openstack(22be8fae-cbe5-4573-af02-07525ab3444b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:50:32 crc kubenswrapper[4959]: E1003 13:50:32.462133 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" podUID="22be8fae-cbe5-4573-af02-07525ab3444b" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504238 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504521 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504540 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504583 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504621 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.504662 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zmsd\" (UniqueName: \"kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.506507 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.508422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.509312 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.512499 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.514777 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.515757 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.516705 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.531007 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zmsd\" (UniqueName: \"kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.544844 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.552177 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerStarted","Data":"8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce"} Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.626877 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:32 crc kubenswrapper[4959]: I1003 13:50:32.930940 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.011088 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.022630 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.120219 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:50:33 crc kubenswrapper[4959]: W1003 13:50:33.641998 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30542159_188f_4508_a1b4_70285f59aef3.slice/crio-5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e WatchSource:0}: Error finding container 5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e: Status 404 returned error can't find the container with id 5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e Oct 03 13:50:33 crc kubenswrapper[4959]: W1003 13:50:33.644439 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb9f7e3a_3cea_4bfd_986b_3a026d04075f.slice/crio-6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de WatchSource:0}: Error finding container 6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de: Status 404 returned error can't find the container with id 6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de Oct 03 13:50:33 crc kubenswrapper[4959]: W1003 13:50:33.647454 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bc8649d_27d5_4834_bc5d_d46f32cbdd0a.slice/crio-fc03974bba56fdd16f826294c190b9492cd471046c9044639a23747af62a869c WatchSource:0}: Error finding container fc03974bba56fdd16f826294c190b9492cd471046c9044639a23747af62a869c: Status 404 returned error can't find the container with id fc03974bba56fdd16f826294c190b9492cd471046c9044639a23747af62a869c Oct 03 13:50:33 crc kubenswrapper[4959]: W1003 13:50:33.657464 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ceadbac_10d3_476b_8250_5005c4391e71.slice/crio-46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c WatchSource:0}: Error finding container 46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c: Status 404 returned error can't find the container with id 46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.709635 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.717281 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.826222 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config\") pod \"ab295ee6-3a3d-449a-9bc5-9176cad79051\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.826286 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config\") pod \"22be8fae-cbe5-4573-af02-07525ab3444b\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.826346 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr5nw\" (UniqueName: \"kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw\") pod \"22be8fae-cbe5-4573-af02-07525ab3444b\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.826412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc\") pod \"22be8fae-cbe5-4573-af02-07525ab3444b\" (UID: \"22be8fae-cbe5-4573-af02-07525ab3444b\") " Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.826509 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98h7k\" (UniqueName: \"kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k\") pod \"ab295ee6-3a3d-449a-9bc5-9176cad79051\" (UID: \"ab295ee6-3a3d-449a-9bc5-9176cad79051\") " Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.827024 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config" (OuterVolumeSpecName: "config") pod "22be8fae-cbe5-4573-af02-07525ab3444b" (UID: "22be8fae-cbe5-4573-af02-07525ab3444b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.827079 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config" (OuterVolumeSpecName: "config") pod "ab295ee6-3a3d-449a-9bc5-9176cad79051" (UID: "ab295ee6-3a3d-449a-9bc5-9176cad79051"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.827933 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "22be8fae-cbe5-4573-af02-07525ab3444b" (UID: "22be8fae-cbe5-4573-af02-07525ab3444b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.830966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw" (OuterVolumeSpecName: "kube-api-access-sr5nw") pod "22be8fae-cbe5-4573-af02-07525ab3444b" (UID: "22be8fae-cbe5-4573-af02-07525ab3444b"). InnerVolumeSpecName "kube-api-access-sr5nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.832158 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k" (OuterVolumeSpecName: "kube-api-access-98h7k") pod "ab295ee6-3a3d-449a-9bc5-9176cad79051" (UID: "ab295ee6-3a3d-449a-9bc5-9176cad79051"). InnerVolumeSpecName "kube-api-access-98h7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.927900 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.928234 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr5nw\" (UniqueName: \"kubernetes.io/projected/22be8fae-cbe5-4573-af02-07525ab3444b-kube-api-access-sr5nw\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.928248 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22be8fae-cbe5-4573-af02-07525ab3444b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.928259 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98h7k\" (UniqueName: \"kubernetes.io/projected/ab295ee6-3a3d-449a-9bc5-9176cad79051-kube-api-access-98h7k\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:33 crc kubenswrapper[4959]: I1003 13:50:33.928269 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab295ee6-3a3d-449a-9bc5-9176cad79051-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.123035 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:50:34 crc kubenswrapper[4959]: W1003 13:50:34.127622 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8729e6fd_0547_43a9_bbef_d4dcbd7475bc.slice/crio-61e7020ede208fc0608d9a051b91870568fedaded24984255e5a778a1f2461e6 WatchSource:0}: Error finding container 61e7020ede208fc0608d9a051b91870568fedaded24984255e5a778a1f2461e6: Status 404 returned error can't find the container with id 61e7020ede208fc0608d9a051b91870568fedaded24984255e5a778a1f2461e6 Oct 03 13:50:34 crc kubenswrapper[4959]: W1003 13:50:34.233773 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod303c0421_debb_437b_8a12_4ff587f419f3.slice/crio-079309bc50bf499106e77138fdf5d99d3c5cd83c60ecdcc55627bfead97d9b11 WatchSource:0}: Error finding container 079309bc50bf499106e77138fdf5d99d3c5cd83c60ecdcc55627bfead97d9b11: Status 404 returned error can't find the container with id 079309bc50bf499106e77138fdf5d99d3c5cd83c60ecdcc55627bfead97d9b11 Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.235600 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.329916 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.416169 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:50:34 crc kubenswrapper[4959]: W1003 13:50:34.478223 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dbba504_e66c_4dbe_abe3_4e69ceec1a23.slice/crio-2594e4b02494d055c82a0fcaa360f604167c2d72ae3b70140a42cc2607d7e46c WatchSource:0}: Error finding container 2594e4b02494d055c82a0fcaa360f604167c2d72ae3b70140a42cc2607d7e46c: Status 404 returned error can't find the container with id 2594e4b02494d055c82a0fcaa360f604167c2d72ae3b70140a42cc2607d7e46c Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.567073 4959 generic.go:334] "Generic (PLEG): container finished" podID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerID="3f95ee24d6c6b8e4ad4c10ddf0efe083656e7139ed31868e3fe196b95a7d1ad6" exitCode=0 Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.567151 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" event={"ID":"2613de27-74fb-444d-b431-7c083c7fc5b3","Type":"ContainerDied","Data":"3f95ee24d6c6b8e4ad4c10ddf0efe083656e7139ed31868e3fe196b95a7d1ad6"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.568343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a","Type":"ContainerStarted","Data":"fc03974bba56fdd16f826294c190b9492cd471046c9044639a23747af62a869c"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.570856 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh" event={"ID":"8729e6fd-0547-43a9-bbef-d4dcbd7475bc","Type":"ContainerStarted","Data":"61e7020ede208fc0608d9a051b91870568fedaded24984255e5a778a1f2461e6"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.572056 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerStarted","Data":"6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.574312 4959 generic.go:334] "Generic (PLEG): container finished" podID="cb452d4d-d280-4035-9a47-d240d9051982" containerID="ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2" exitCode=0 Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.574529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" event={"ID":"cb452d4d-d280-4035-9a47-d240d9051982","Type":"ContainerDied","Data":"ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.576763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerStarted","Data":"2594e4b02494d055c82a0fcaa360f604167c2d72ae3b70140a42cc2607d7e46c"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.579519 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerStarted","Data":"4e339b80df0668dff924d38130d42e6972293d3ab0b65a87d13a033ae0d1d964"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.587594 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerStarted","Data":"46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.594531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"30542159-188f-4508-a1b4-70285f59aef3","Type":"ContainerStarted","Data":"5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.595987 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" event={"ID":"22be8fae-cbe5-4573-af02-07525ab3444b","Type":"ContainerDied","Data":"f337cbfdbd43d839aebfec1fa2a5255461bbd7342fcda37d0b286ea8d169a6b1"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.596061 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gx9x4" Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.600327 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerStarted","Data":"079309bc50bf499106e77138fdf5d99d3c5cd83c60ecdcc55627bfead97d9b11"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.602059 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" event={"ID":"ab295ee6-3a3d-449a-9bc5-9176cad79051","Type":"ContainerDied","Data":"a350d82a55f8472b3fba347e8154050c8306e9121229ea02940531a5df424503"} Oct 03 13:50:34 crc kubenswrapper[4959]: I1003 13:50:34.602124 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wtv4p" Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.008711 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.016410 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gx9x4"] Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.031473 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.036459 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wtv4p"] Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.612174 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" event={"ID":"cb452d4d-d280-4035-9a47-d240d9051982","Type":"ContainerStarted","Data":"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f"} Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.612473 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.614500 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerStarted","Data":"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307"} Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.617313 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerStarted","Data":"1e12c25b84bbfc17840c8beb88edad3a093b4fd3bcbb1e87e14c1aaf4a2ed572"} Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.633305 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" podStartSLOduration=3.762012279 podStartE2EDuration="17.633291638s" podCreationTimestamp="2025-10-03 13:50:18 +0000 UTC" firstStartedPulling="2025-10-03 13:50:19.890989433 +0000 UTC m=+1189.094332850" lastFinishedPulling="2025-10-03 13:50:33.762268792 +0000 UTC m=+1202.965612209" observedRunningTime="2025-10-03 13:50:35.629215529 +0000 UTC m=+1204.832558946" watchObservedRunningTime="2025-10-03 13:50:35.633291638 +0000 UTC m=+1204.836635055" Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.696307 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22be8fae-cbe5-4573-af02-07525ab3444b" path="/var/lib/kubelet/pods/22be8fae-cbe5-4573-af02-07525ab3444b/volumes" Oct 03 13:50:35 crc kubenswrapper[4959]: I1003 13:50:35.696733 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab295ee6-3a3d-449a-9bc5-9176cad79051" path="/var/lib/kubelet/pods/ab295ee6-3a3d-449a-9bc5-9176cad79051/volumes" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.660222 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"30542159-188f-4508-a1b4-70285f59aef3","Type":"ContainerStarted","Data":"aae8036f06d6bb33643dcc5d309e687768c23fb0437205d1473dc1f70e1e4751"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.660821 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.662569 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh" event={"ID":"8729e6fd-0547-43a9-bbef-d4dcbd7475bc","Type":"ContainerStarted","Data":"d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.662662 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-k4vvh" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.664772 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerStarted","Data":"c979a79ee658ed339b2913d7a6893fbe42a93bd9add0d47862376c97d3845ac0"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.666875 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerStarted","Data":"0854357ab58301987a5e872493993a441399ef76425c0597c5a86e12bd783410"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.668125 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerStarted","Data":"363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.669944 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerStarted","Data":"97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.672487 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" event={"ID":"2613de27-74fb-444d-b431-7c083c7fc5b3","Type":"ContainerStarted","Data":"fbfc62b1ba99691eccbc3fe4ae618e2b7e849f29c9ec5b95dd610a439f06b52b"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.672668 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.673836 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a","Type":"ContainerStarted","Data":"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.673986 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.675327 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerStarted","Data":"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e"} Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.682394 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.15058132 podStartE2EDuration="19.682379704s" podCreationTimestamp="2025-10-03 13:50:22 +0000 UTC" firstStartedPulling="2025-10-03 13:50:33.651313507 +0000 UTC m=+1202.854656924" lastFinishedPulling="2025-10-03 13:50:40.183111891 +0000 UTC m=+1209.386455308" observedRunningTime="2025-10-03 13:50:41.678717686 +0000 UTC m=+1210.882061103" watchObservedRunningTime="2025-10-03 13:50:41.682379704 +0000 UTC m=+1210.885723121" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.709658 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" podStartSLOduration=9.105834663 podStartE2EDuration="23.709639261s" podCreationTimestamp="2025-10-03 13:50:18 +0000 UTC" firstStartedPulling="2025-10-03 13:50:19.116883222 +0000 UTC m=+1188.320226649" lastFinishedPulling="2025-10-03 13:50:33.72068783 +0000 UTC m=+1202.924031247" observedRunningTime="2025-10-03 13:50:41.702577981 +0000 UTC m=+1210.905921398" watchObservedRunningTime="2025-10-03 13:50:41.709639261 +0000 UTC m=+1210.912982668" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.723085 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-k4vvh" podStartSLOduration=8.451787841 podStartE2EDuration="14.723066345s" podCreationTimestamp="2025-10-03 13:50:27 +0000 UTC" firstStartedPulling="2025-10-03 13:50:34.130693603 +0000 UTC m=+1203.334037030" lastFinishedPulling="2025-10-03 13:50:40.401972107 +0000 UTC m=+1209.605315534" observedRunningTime="2025-10-03 13:50:41.721830095 +0000 UTC m=+1210.925173512" watchObservedRunningTime="2025-10-03 13:50:41.723066345 +0000 UTC m=+1210.926409752" Oct 03 13:50:41 crc kubenswrapper[4959]: I1003 13:50:41.785823 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.090999017 podStartE2EDuration="17.785805438s" podCreationTimestamp="2025-10-03 13:50:24 +0000 UTC" firstStartedPulling="2025-10-03 13:50:33.651344188 +0000 UTC m=+1202.854687605" lastFinishedPulling="2025-10-03 13:50:41.346150599 +0000 UTC m=+1210.549494026" observedRunningTime="2025-10-03 13:50:41.780037969 +0000 UTC m=+1210.983381386" watchObservedRunningTime="2025-10-03 13:50:41.785805438 +0000 UTC m=+1210.989148855" Oct 03 13:50:42 crc kubenswrapper[4959]: I1003 13:50:42.688630 4959 generic.go:334] "Generic (PLEG): container finished" podID="303c0421-debb-437b-8a12-4ff587f419f3" containerID="354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e" exitCode=0 Oct 03 13:50:42 crc kubenswrapper[4959]: I1003 13:50:42.690007 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerDied","Data":"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e"} Oct 03 13:50:43 crc kubenswrapper[4959]: I1003 13:50:43.699294 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerStarted","Data":"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb"} Oct 03 13:50:44 crc kubenswrapper[4959]: I1003 13:50:44.341326 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:50:44 crc kubenswrapper[4959]: I1003 13:50:44.396483 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:44 crc kubenswrapper[4959]: I1003 13:50:44.396682 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="dnsmasq-dns" containerID="cri-o://fbfc62b1ba99691eccbc3fe4ae618e2b7e849f29c9ec5b95dd610a439f06b52b" gracePeriod=10 Oct 03 13:50:44 crc kubenswrapper[4959]: I1003 13:50:44.740592 4959 generic.go:334] "Generic (PLEG): container finished" podID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerID="fbfc62b1ba99691eccbc3fe4ae618e2b7e849f29c9ec5b95dd610a439f06b52b" exitCode=0 Oct 03 13:50:44 crc kubenswrapper[4959]: I1003 13:50:44.740664 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" event={"ID":"2613de27-74fb-444d-b431-7c083c7fc5b3","Type":"ContainerDied","Data":"fbfc62b1ba99691eccbc3fe4ae618e2b7e849f29c9ec5b95dd610a439f06b52b"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.373333 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.532526 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc\") pod \"2613de27-74fb-444d-b431-7c083c7fc5b3\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.532816 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config\") pod \"2613de27-74fb-444d-b431-7c083c7fc5b3\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.532911 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d4bw\" (UniqueName: \"kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw\") pod \"2613de27-74fb-444d-b431-7c083c7fc5b3\" (UID: \"2613de27-74fb-444d-b431-7c083c7fc5b3\") " Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.535884 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw" (OuterVolumeSpecName: "kube-api-access-6d4bw") pod "2613de27-74fb-444d-b431-7c083c7fc5b3" (UID: "2613de27-74fb-444d-b431-7c083c7fc5b3"). InnerVolumeSpecName "kube-api-access-6d4bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.563007 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config" (OuterVolumeSpecName: "config") pod "2613de27-74fb-444d-b431-7c083c7fc5b3" (UID: "2613de27-74fb-444d-b431-7c083c7fc5b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.567169 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2613de27-74fb-444d-b431-7c083c7fc5b3" (UID: "2613de27-74fb-444d-b431-7c083c7fc5b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.634984 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.635038 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2613de27-74fb-444d-b431-7c083c7fc5b3-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.635052 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d4bw\" (UniqueName: \"kubernetes.io/projected/2613de27-74fb-444d-b431-7c083c7fc5b3-kube-api-access-6d4bw\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.752033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerStarted","Data":"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.752365 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.755954 4959 generic.go:334] "Generic (PLEG): container finished" podID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerID="0854357ab58301987a5e872493993a441399ef76425c0597c5a86e12bd783410" exitCode=0 Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.756029 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerDied","Data":"0854357ab58301987a5e872493993a441399ef76425c0597c5a86e12bd783410"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.759084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerStarted","Data":"eba109515bfe8301850c53df177bc5231193d427a4c857c848ed3c8d12e4e21b"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.761668 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerStarted","Data":"18ea0e8b0cc153504a9c772e75f38038e8891d7353b017f050ebb59b0ecbf766"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.765451 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.765483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cr4bb" event={"ID":"2613de27-74fb-444d-b431-7c083c7fc5b3","Type":"ContainerDied","Data":"020d66a7052bcb9a66ad5464c614925caecc426a4998c0114f9771ac0f4682d5"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.765567 4959 scope.go:117] "RemoveContainer" containerID="fbfc62b1ba99691eccbc3fe4ae618e2b7e849f29c9ec5b95dd610a439f06b52b" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.767452 4959 generic.go:334] "Generic (PLEG): container finished" podID="9ceadbac-10d3-476b-8250-5005c4391e71" containerID="c979a79ee658ed339b2913d7a6893fbe42a93bd9add0d47862376c97d3845ac0" exitCode=0 Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.767519 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerDied","Data":"c979a79ee658ed339b2913d7a6893fbe42a93bd9add0d47862376c97d3845ac0"} Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.785092 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2bhp5" podStartSLOduration=11.529295299 podStartE2EDuration="17.785070659s" podCreationTimestamp="2025-10-03 13:50:28 +0000 UTC" firstStartedPulling="2025-10-03 13:50:34.239331853 +0000 UTC m=+1203.442675280" lastFinishedPulling="2025-10-03 13:50:40.495107223 +0000 UTC m=+1209.698450640" observedRunningTime="2025-10-03 13:50:45.77226302 +0000 UTC m=+1214.975606467" watchObservedRunningTime="2025-10-03 13:50:45.785070659 +0000 UTC m=+1214.988414086" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.804579 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.860244622 podStartE2EDuration="18.804557589s" podCreationTimestamp="2025-10-03 13:50:27 +0000 UTC" firstStartedPulling="2025-10-03 13:50:34.387557456 +0000 UTC m=+1203.590900873" lastFinishedPulling="2025-10-03 13:50:45.331870423 +0000 UTC m=+1214.535213840" observedRunningTime="2025-10-03 13:50:45.799320522 +0000 UTC m=+1215.002663959" watchObservedRunningTime="2025-10-03 13:50:45.804557589 +0000 UTC m=+1215.007901006" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.806761 4959 scope.go:117] "RemoveContainer" containerID="3f95ee24d6c6b8e4ad4c10ddf0efe083656e7139ed31868e3fe196b95a7d1ad6" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.884891 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.025042073 podStartE2EDuration="14.884869734s" podCreationTimestamp="2025-10-03 13:50:31 +0000 UTC" firstStartedPulling="2025-10-03 13:50:34.480719272 +0000 UTC m=+1203.684062729" lastFinishedPulling="2025-10-03 13:50:45.340546973 +0000 UTC m=+1214.543890390" observedRunningTime="2025-10-03 13:50:45.879459135 +0000 UTC m=+1215.082802562" watchObservedRunningTime="2025-10-03 13:50:45.884869734 +0000 UTC m=+1215.088213151" Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.913636 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:45 crc kubenswrapper[4959]: I1003 13:50:45.930533 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cr4bb"] Oct 03 13:50:46 crc kubenswrapper[4959]: I1003 13:50:46.781506 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerStarted","Data":"f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2"} Oct 03 13:50:46 crc kubenswrapper[4959]: I1003 13:50:46.785128 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerStarted","Data":"1af66b255906c8775c33d1bfe9a4fbf4e95006a4472a5b2389c20102dfd422b5"} Oct 03 13:50:46 crc kubenswrapper[4959]: I1003 13:50:46.785341 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:50:46 crc kubenswrapper[4959]: I1003 13:50:46.803656 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.104188788 podStartE2EDuration="26.803640344s" podCreationTimestamp="2025-10-03 13:50:20 +0000 UTC" firstStartedPulling="2025-10-03 13:50:33.687431948 +0000 UTC m=+1202.890775385" lastFinishedPulling="2025-10-03 13:50:40.386883514 +0000 UTC m=+1209.590226941" observedRunningTime="2025-10-03 13:50:46.799127875 +0000 UTC m=+1216.002471292" watchObservedRunningTime="2025-10-03 13:50:46.803640344 +0000 UTC m=+1216.006983761" Oct 03 13:50:46 crc kubenswrapper[4959]: I1003 13:50:46.836524 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.869298546 podStartE2EDuration="25.836495696s" podCreationTimestamp="2025-10-03 13:50:21 +0000 UTC" firstStartedPulling="2025-10-03 13:50:33.651353788 +0000 UTC m=+1202.854697205" lastFinishedPulling="2025-10-03 13:50:40.618550898 +0000 UTC m=+1209.821894355" observedRunningTime="2025-10-03 13:50:46.824876986 +0000 UTC m=+1216.028220403" watchObservedRunningTime="2025-10-03 13:50:46.836495696 +0000 UTC m=+1216.039839143" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.255616 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.298461 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.627570 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.627633 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.661593 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.694264 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" path="/var/lib/kubelet/pods/2613de27-74fb-444d-b431-7c083c7fc5b3/volumes" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.792721 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.914379 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 03 13:50:47 crc kubenswrapper[4959]: I1003 13:50:47.920105 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.101373 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.187440 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tqgvv"] Oct 03 13:50:48 crc kubenswrapper[4959]: E1003 13:50:48.188453 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="init" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.188477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="init" Oct 03 13:50:48 crc kubenswrapper[4959]: E1003 13:50:48.188491 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="dnsmasq-dns" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.188499 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="dnsmasq-dns" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.188879 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2613de27-74fb-444d-b431-7c083c7fc5b3" containerName="dnsmasq-dns" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.190351 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.203137 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.234846 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tqgvv"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.250365 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.252871 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.265820 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.266915 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.279958 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.280017 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.280284 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.280480 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnp9c\" (UniqueName: \"kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.358627 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tqgvv"] Oct 03 13:50:48 crc kubenswrapper[4959]: E1003 13:50:48.359274 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-rnp9c ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" podUID="05655f8f-39f3-4928-a23a-404e99a92e05" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.377920 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.379226 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.383771 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.383981 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wjlsc" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.384154 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.384391 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386220 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnp9c\" (UniqueName: \"kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386299 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386341 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386393 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw968\" (UniqueName: \"kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386547 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.386574 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.387699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.388660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.388972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.397871 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.399524 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.403450 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.418027 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnp9c\" (UniqueName: \"kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c\") pod \"dnsmasq-dns-7fd796d7df-tqgvv\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.428515 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.464539 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490325 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490374 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490406 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490423 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490538 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490560 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg86x\" (UniqueName: \"kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490581 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490598 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw968\" (UniqueName: \"kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490653 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490673 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490702 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxksn\" (UniqueName: \"kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.490728 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.491340 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.491431 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.491487 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.495578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.497290 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.516284 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw968\" (UniqueName: \"kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968\") pod \"ovn-controller-metrics-kx2cq\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.569012 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.592649 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.592713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.592741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593712 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593734 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg86x\" (UniqueName: \"kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593760 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593786 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593809 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593830 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxksn\" (UniqueName: \"kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.593902 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.594127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.594562 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.594599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.594667 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.595330 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.595663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.595866 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.598645 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.601315 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.613551 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg86x\" (UniqueName: \"kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.614090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.623372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxksn\" (UniqueName: \"kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn\") pod \"dnsmasq-dns-86db49b7ff-7kdgk\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.707723 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.764753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.814369 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.837682 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.899157 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc\") pod \"05655f8f-39f3-4928-a23a-404e99a92e05\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.899394 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config\") pod \"05655f8f-39f3-4928-a23a-404e99a92e05\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.899451 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnp9c\" (UniqueName: \"kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c\") pod \"05655f8f-39f3-4928-a23a-404e99a92e05\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.899477 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb\") pod \"05655f8f-39f3-4928-a23a-404e99a92e05\" (UID: \"05655f8f-39f3-4928-a23a-404e99a92e05\") " Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.900250 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "05655f8f-39f3-4928-a23a-404e99a92e05" (UID: "05655f8f-39f3-4928-a23a-404e99a92e05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.900276 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config" (OuterVolumeSpecName: "config") pod "05655f8f-39f3-4928-a23a-404e99a92e05" (UID: "05655f8f-39f3-4928-a23a-404e99a92e05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.900635 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05655f8f-39f3-4928-a23a-404e99a92e05" (UID: "05655f8f-39f3-4928-a23a-404e99a92e05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.901421 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.901445 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.901457 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05655f8f-39f3-4928-a23a-404e99a92e05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:48 crc kubenswrapper[4959]: I1003 13:50:48.909509 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c" (OuterVolumeSpecName: "kube-api-access-rnp9c") pod "05655f8f-39f3-4928-a23a-404e99a92e05" (UID: "05655f8f-39f3-4928-a23a-404e99a92e05"). InnerVolumeSpecName "kube-api-access-rnp9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.002744 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnp9c\" (UniqueName: \"kubernetes.io/projected/05655f8f-39f3-4928-a23a-404e99a92e05-kube-api-access-rnp9c\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.062607 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:50:49 crc kubenswrapper[4959]: W1003 13:50:49.067307 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bfc789_a017_43e7_ba4f_96b93514bcf4.slice/crio-bf8ac294982419fa3eb8f1a46d0985ec703ea43985cbf877285ce1bc71cb906b WatchSource:0}: Error finding container bf8ac294982419fa3eb8f1a46d0985ec703ea43985cbf877285ce1bc71cb906b: Status 404 returned error can't find the container with id bf8ac294982419fa3eb8f1a46d0985ec703ea43985cbf877285ce1bc71cb906b Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.181010 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.279876 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.822608 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" event={"ID":"f1dd2167-e90b-40db-a025-5fa92a428897","Type":"ContainerStarted","Data":"d2b64f0eaf6cacefe33a360b156f16d6631641a6623a93483e01bfa3cc05a059"} Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.825065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kx2cq" event={"ID":"a4bfc789-a017-43e7-ba4f-96b93514bcf4","Type":"ContainerStarted","Data":"bf8ac294982419fa3eb8f1a46d0985ec703ea43985cbf877285ce1bc71cb906b"} Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.826883 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tqgvv" Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.826969 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerStarted","Data":"d25d5d7ca31fbf65ba57aee081929fb858eb137519251181c4cf138741675524"} Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.860137 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tqgvv"] Oct 03 13:50:49 crc kubenswrapper[4959]: I1003 13:50:49.880764 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tqgvv"] Oct 03 13:50:51 crc kubenswrapper[4959]: I1003 13:50:51.388620 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 03 13:50:51 crc kubenswrapper[4959]: I1003 13:50:51.390944 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 03 13:50:51 crc kubenswrapper[4959]: I1003 13:50:51.694084 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05655f8f-39f3-4928-a23a-404e99a92e05" path="/var/lib/kubelet/pods/05655f8f-39f3-4928-a23a-404e99a92e05/volumes" Oct 03 13:50:52 crc kubenswrapper[4959]: I1003 13:50:52.768765 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:52 crc kubenswrapper[4959]: I1003 13:50:52.770286 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:53 crc kubenswrapper[4959]: I1003 13:50:53.102215 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 03 13:50:53 crc kubenswrapper[4959]: I1003 13:50:53.151686 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" probeResult="failure" output=< Oct 03 13:50:53 crc kubenswrapper[4959]: wsrep_local_state_comment (Joined) differs from Synced Oct 03 13:50:53 crc kubenswrapper[4959]: > Oct 03 13:50:54 crc kubenswrapper[4959]: I1003 13:50:54.851553 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 13:50:54 crc kubenswrapper[4959]: I1003 13:50:54.895367 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:54 crc kubenswrapper[4959]: I1003 13:50:54.924573 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:50:54 crc kubenswrapper[4959]: I1003 13:50:54.926601 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:54 crc kubenswrapper[4959]: I1003 13:50:54.937999 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.031417 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pq7c\" (UniqueName: \"kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.031496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.031753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.031875 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.031904 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.133504 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.133555 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.133589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pq7c\" (UniqueName: \"kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.133629 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.133716 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.134537 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.134550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.134612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.134627 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.157086 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pq7c\" (UniqueName: \"kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c\") pod \"dnsmasq-dns-698758b865-7lrmq\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.254431 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:55 crc kubenswrapper[4959]: W1003 13:50:55.689459 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod593147fe_5977_46ff_8f90_627e3a48142c.slice/crio-483a376d93b8ec0091ddaaf53035128b8f2ef19e92e59348ac63a5c1a313f3e4 WatchSource:0}: Error finding container 483a376d93b8ec0091ddaaf53035128b8f2ef19e92e59348ac63a5c1a313f3e4: Status 404 returned error can't find the container with id 483a376d93b8ec0091ddaaf53035128b8f2ef19e92e59348ac63a5c1a313f3e4 Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.694814 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:50:55 crc kubenswrapper[4959]: I1003 13:50:55.872615 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7lrmq" event={"ID":"593147fe-5977-46ff-8f90-627e3a48142c","Type":"ContainerStarted","Data":"483a376d93b8ec0091ddaaf53035128b8f2ef19e92e59348ac63a5c1a313f3e4"} Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.043271 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.048488 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.050942 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.050976 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.051354 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6mmv4" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.051564 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.071025 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.151504 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.151564 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.151663 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.151684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.151706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42p2c\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.252617 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.252661 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.252682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42p2c\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.252750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.252769 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.252904 4959 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.252974 4959 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.253048 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift podName:29a02381-a818-4890-9c9a-089fc14e565d nodeName:}" failed. No retries permitted until 2025-10-03 13:50:56.753025983 +0000 UTC m=+1225.956369470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift") pod "swift-storage-0" (UID: "29a02381-a818-4890-9c9a-089fc14e565d") : configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.253098 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.253240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.253312 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.273754 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.292083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42p2c\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.536149 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-92xpj"] Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.537160 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.539363 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.545560 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.545596 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.546563 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-92xpj"] Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.662280 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.662381 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.662600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.662709 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4cdj\" (UniqueName: \"kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.662866 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.663005 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.663145 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765297 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765448 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765531 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4cdj\" (UniqueName: \"kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765582 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765628 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.765657 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.766040 4959 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.766065 4959 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: E1003 13:50:56.766118 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift podName:29a02381-a818-4890-9c9a-089fc14e565d nodeName:}" failed. No retries permitted until 2025-10-03 13:50:57.766100982 +0000 UTC m=+1226.969444399 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift") pod "swift-storage-0" (UID: "29a02381-a818-4890-9c9a-089fc14e565d") : configmap "swift-ring-files" not found Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.766286 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.766322 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.766587 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.769752 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.769840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.770287 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.781601 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4cdj\" (UniqueName: \"kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj\") pod \"swift-ring-rebalance-92xpj\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.857274 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.885852 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" event={"ID":"f1dd2167-e90b-40db-a025-5fa92a428897","Type":"ContainerStarted","Data":"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb"} Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.893771 4959 generic.go:334] "Generic (PLEG): container finished" podID="593147fe-5977-46ff-8f90-627e3a48142c" containerID="13e2b340838fa505ce2c1b982d466277a68ebe4aaad9e667134be02bc57d0a38" exitCode=0 Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.893828 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7lrmq" event={"ID":"593147fe-5977-46ff-8f90-627e3a48142c","Type":"ContainerDied","Data":"13e2b340838fa505ce2c1b982d466277a68ebe4aaad9e667134be02bc57d0a38"} Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.913716 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kx2cq" event={"ID":"a4bfc789-a017-43e7-ba4f-96b93514bcf4","Type":"ContainerStarted","Data":"ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5"} Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.953129 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kx2cq" podStartSLOduration=8.95310965 podStartE2EDuration="8.95310965s" podCreationTimestamp="2025-10-03 13:50:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:50:56.94688993 +0000 UTC m=+1226.150233347" watchObservedRunningTime="2025-10-03 13:50:56.95310965 +0000 UTC m=+1226.156453067" Oct 03 13:50:56 crc kubenswrapper[4959]: I1003 13:50:56.966487 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.036117 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.465945 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.528168 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-92xpj"] Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.584711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.584838 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.584898 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxksn\" (UniqueName: \"kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.584929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.584977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.590247 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn" (OuterVolumeSpecName: "kube-api-access-dxksn") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897"). InnerVolumeSpecName "kube-api-access-dxksn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.613630 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:57 crc kubenswrapper[4959]: E1003 13:50:57.622885 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config podName:f1dd2167-e90b-40db-a025-5fa92a428897 nodeName:}" failed. No retries permitted until 2025-10-03 13:50:58.122847846 +0000 UTC m=+1227.326191263 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897") : error deleting /var/lib/kubelet/pods/f1dd2167-e90b-40db-a025-5fa92a428897/volume-subpaths: remove /var/lib/kubelet/pods/f1dd2167-e90b-40db-a025-5fa92a428897/volume-subpaths: no such file or directory Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.622904 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.622966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.687484 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxksn\" (UniqueName: \"kubernetes.io/projected/f1dd2167-e90b-40db-a025-5fa92a428897-kube-api-access-dxksn\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.687526 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.687538 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.687546 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.788713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:57 crc kubenswrapper[4959]: E1003 13:50:57.788847 4959 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:50:57 crc kubenswrapper[4959]: E1003 13:50:57.788866 4959 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:50:57 crc kubenswrapper[4959]: E1003 13:50:57.788913 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift podName:29a02381-a818-4890-9c9a-089fc14e565d nodeName:}" failed. No retries permitted until 2025-10-03 13:50:59.788898918 +0000 UTC m=+1228.992242335 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift") pod "swift-storage-0" (UID: "29a02381-a818-4890-9c9a-089fc14e565d") : configmap "swift-ring-files" not found Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.921250 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7lrmq" event={"ID":"593147fe-5977-46ff-8f90-627e3a48142c","Type":"ContainerStarted","Data":"7bd9b89a81b4425bd0f9e9824cc5249a65423d9d48ba6caf70db91ea28349056"} Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.921385 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.922671 4959 generic.go:334] "Generic (PLEG): container finished" podID="f1dd2167-e90b-40db-a025-5fa92a428897" containerID="5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb" exitCode=0 Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.922714 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.922752 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" event={"ID":"f1dd2167-e90b-40db-a025-5fa92a428897","Type":"ContainerDied","Data":"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb"} Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.922787 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7kdgk" event={"ID":"f1dd2167-e90b-40db-a025-5fa92a428897","Type":"ContainerDied","Data":"d2b64f0eaf6cacefe33a360b156f16d6631641a6623a93483e01bfa3cc05a059"} Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.922811 4959 scope.go:117] "RemoveContainer" containerID="5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.924925 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-92xpj" event={"ID":"9655f477-fe02-412d-b52d-6aea687e1e57","Type":"ContainerStarted","Data":"774fd1d79d4cadf2dfb5b1a7537f3510f284a3ce3107f456b21eb58a82e4fb52"} Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.943920 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-7lrmq" podStartSLOduration=3.943903016 podStartE2EDuration="3.943903016s" podCreationTimestamp="2025-10-03 13:50:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:50:57.938588477 +0000 UTC m=+1227.141931904" watchObservedRunningTime="2025-10-03 13:50:57.943903016 +0000 UTC m=+1227.147246433" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.955428 4959 scope.go:117] "RemoveContainer" containerID="5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb" Oct 03 13:50:57 crc kubenswrapper[4959]: E1003 13:50:57.955888 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb\": container with ID starting with 5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb not found: ID does not exist" containerID="5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb" Oct 03 13:50:57 crc kubenswrapper[4959]: I1003 13:50:57.955942 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb"} err="failed to get container status \"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb\": rpc error: code = NotFound desc = could not find container \"5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb\": container with ID starting with 5353ca07573e8f0132cd4e3dd23a564b1552a6badb2f8ad8250cd71a210b7bbb not found: ID does not exist" Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.198804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") pod \"f1dd2167-e90b-40db-a025-5fa92a428897\" (UID: \"f1dd2167-e90b-40db-a025-5fa92a428897\") " Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.199540 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config" (OuterVolumeSpecName: "config") pod "f1dd2167-e90b-40db-a025-5fa92a428897" (UID: "f1dd2167-e90b-40db-a025-5fa92a428897"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.199780 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1dd2167-e90b-40db-a025-5fa92a428897-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.288209 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.293804 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7kdgk"] Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.937028 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerStarted","Data":"9a443e77a12743b8133a947d16386d9a7160da93b531ed00b22855dabcff87e6"} Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.937378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerStarted","Data":"7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e"} Oct 03 13:50:58 crc kubenswrapper[4959]: I1003 13:50:58.959481 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.292373549 podStartE2EDuration="10.959464069s" podCreationTimestamp="2025-10-03 13:50:48 +0000 UTC" firstStartedPulling="2025-10-03 13:50:49.188890856 +0000 UTC m=+1218.392234313" lastFinishedPulling="2025-10-03 13:50:57.855981416 +0000 UTC m=+1227.059324833" observedRunningTime="2025-10-03 13:50:58.955698018 +0000 UTC m=+1228.159041435" watchObservedRunningTime="2025-10-03 13:50:58.959464069 +0000 UTC m=+1228.162807486" Oct 03 13:50:59 crc kubenswrapper[4959]: I1003 13:50:59.696120 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1dd2167-e90b-40db-a025-5fa92a428897" path="/var/lib/kubelet/pods/f1dd2167-e90b-40db-a025-5fa92a428897/volumes" Oct 03 13:50:59 crc kubenswrapper[4959]: I1003 13:50:59.825631 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:50:59 crc kubenswrapper[4959]: E1003 13:50:59.825797 4959 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:50:59 crc kubenswrapper[4959]: E1003 13:50:59.825812 4959 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:50:59 crc kubenswrapper[4959]: E1003 13:50:59.825852 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift podName:29a02381-a818-4890-9c9a-089fc14e565d nodeName:}" failed. No retries permitted until 2025-10-03 13:51:03.825838184 +0000 UTC m=+1233.029181601 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift") pod "swift-storage-0" (UID: "29a02381-a818-4890-9c9a-089fc14e565d") : configmap "swift-ring-files" not found Oct 03 13:50:59 crc kubenswrapper[4959]: I1003 13:50:59.944062 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 03 13:51:00 crc kubenswrapper[4959]: I1003 13:51:00.952666 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-92xpj" event={"ID":"9655f477-fe02-412d-b52d-6aea687e1e57","Type":"ContainerStarted","Data":"956042f00860508daaea28b26ecf1fd459cedf0fd52cb55d9ba130ca282f8dbc"} Oct 03 13:51:01 crc kubenswrapper[4959]: I1003 13:51:01.438985 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 03 13:51:01 crc kubenswrapper[4959]: I1003 13:51:01.984010 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-92xpj" podStartSLOduration=2.9141640840000003 podStartE2EDuration="5.983981169s" podCreationTimestamp="2025-10-03 13:50:56 +0000 UTC" firstStartedPulling="2025-10-03 13:50:57.546235039 +0000 UTC m=+1226.749578456" lastFinishedPulling="2025-10-03 13:51:00.616052104 +0000 UTC m=+1229.819395541" observedRunningTime="2025-10-03 13:51:01.977496481 +0000 UTC m=+1231.180839898" watchObservedRunningTime="2025-10-03 13:51:01.983981169 +0000 UTC m=+1231.187324626" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.763555 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2pszr"] Oct 03 13:51:02 crc kubenswrapper[4959]: E1003 13:51:02.764399 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1dd2167-e90b-40db-a025-5fa92a428897" containerName="init" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.764422 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1dd2167-e90b-40db-a025-5fa92a428897" containerName="init" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.764637 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1dd2167-e90b-40db-a025-5fa92a428897" containerName="init" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.765919 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.771417 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2pszr"] Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.869808 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9jq4\" (UniqueName: \"kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4\") pod \"keystone-db-create-2pszr\" (UID: \"6d640be9-09ac-44ce-af77-68ea96aa6483\") " pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.971104 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9jq4\" (UniqueName: \"kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4\") pod \"keystone-db-create-2pszr\" (UID: \"6d640be9-09ac-44ce-af77-68ea96aa6483\") " pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.972074 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-949rt"] Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.973386 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-949rt" Oct 03 13:51:02 crc kubenswrapper[4959]: I1003 13:51:02.981321 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-949rt"] Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.003204 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9jq4\" (UniqueName: \"kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4\") pod \"keystone-db-create-2pszr\" (UID: \"6d640be9-09ac-44ce-af77-68ea96aa6483\") " pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.073115 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms5dh\" (UniqueName: \"kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh\") pod \"placement-db-create-949rt\" (UID: \"56a3ee2c-e196-4d8d-a796-507963138289\") " pod="openstack/placement-db-create-949rt" Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.088602 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.180013 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms5dh\" (UniqueName: \"kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh\") pod \"placement-db-create-949rt\" (UID: \"56a3ee2c-e196-4d8d-a796-507963138289\") " pod="openstack/placement-db-create-949rt" Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.201501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms5dh\" (UniqueName: \"kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh\") pod \"placement-db-create-949rt\" (UID: \"56a3ee2c-e196-4d8d-a796-507963138289\") " pod="openstack/placement-db-create-949rt" Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.367536 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-949rt" Oct 03 13:51:03 crc kubenswrapper[4959]: W1003 13:51:03.511129 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d640be9_09ac_44ce_af77_68ea96aa6483.slice/crio-04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825 WatchSource:0}: Error finding container 04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825: Status 404 returned error can't find the container with id 04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825 Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.517839 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2pszr"] Oct 03 13:51:03 crc kubenswrapper[4959]: W1003 13:51:03.819125 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56a3ee2c_e196_4d8d_a796_507963138289.slice/crio-5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b WatchSource:0}: Error finding container 5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b: Status 404 returned error can't find the container with id 5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.819887 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-949rt"] Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.893364 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:51:03 crc kubenswrapper[4959]: E1003 13:51:03.893721 4959 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:51:03 crc kubenswrapper[4959]: E1003 13:51:03.893781 4959 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:51:03 crc kubenswrapper[4959]: E1003 13:51:03.893873 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift podName:29a02381-a818-4890-9c9a-089fc14e565d nodeName:}" failed. No retries permitted until 2025-10-03 13:51:11.893845316 +0000 UTC m=+1241.097188763 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift") pod "swift-storage-0" (UID: "29a02381-a818-4890-9c9a-089fc14e565d") : configmap "swift-ring-files" not found Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.978588 4959 generic.go:334] "Generic (PLEG): container finished" podID="6d640be9-09ac-44ce-af77-68ea96aa6483" containerID="b198669f36a1928334af1cd79e08fedc8aebaee6e9b73477d066ad5c78372a5a" exitCode=0 Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.978703 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2pszr" event={"ID":"6d640be9-09ac-44ce-af77-68ea96aa6483","Type":"ContainerDied","Data":"b198669f36a1928334af1cd79e08fedc8aebaee6e9b73477d066ad5c78372a5a"} Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.978735 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2pszr" event={"ID":"6d640be9-09ac-44ce-af77-68ea96aa6483","Type":"ContainerStarted","Data":"04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825"} Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.980378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-949rt" event={"ID":"56a3ee2c-e196-4d8d-a796-507963138289","Type":"ContainerStarted","Data":"998d4944dee5b32d1a16f943f653f08e894041495c582824790d2bdc1559922a"} Oct 03 13:51:03 crc kubenswrapper[4959]: I1003 13:51:03.980410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-949rt" event={"ID":"56a3ee2c-e196-4d8d-a796-507963138289","Type":"ContainerStarted","Data":"5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b"} Oct 03 13:51:04 crc kubenswrapper[4959]: I1003 13:51:04.009508 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-949rt" podStartSLOduration=2.009487261 podStartE2EDuration="2.009487261s" podCreationTimestamp="2025-10-03 13:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:04.00899244 +0000 UTC m=+1233.212335857" watchObservedRunningTime="2025-10-03 13:51:04.009487261 +0000 UTC m=+1233.212830678" Oct 03 13:51:04 crc kubenswrapper[4959]: I1003 13:51:04.986634 4959 generic.go:334] "Generic (PLEG): container finished" podID="56a3ee2c-e196-4d8d-a796-507963138289" containerID="998d4944dee5b32d1a16f943f653f08e894041495c582824790d2bdc1559922a" exitCode=0 Oct 03 13:51:04 crc kubenswrapper[4959]: I1003 13:51:04.986831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-949rt" event={"ID":"56a3ee2c-e196-4d8d-a796-507963138289","Type":"ContainerDied","Data":"998d4944dee5b32d1a16f943f653f08e894041495c582824790d2bdc1559922a"} Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.256420 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.311286 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.311531 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="dnsmasq-dns" containerID="cri-o://991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f" gracePeriod=10 Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.415203 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.528470 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9jq4\" (UniqueName: \"kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4\") pod \"6d640be9-09ac-44ce-af77-68ea96aa6483\" (UID: \"6d640be9-09ac-44ce-af77-68ea96aa6483\") " Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.550349 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4" (OuterVolumeSpecName: "kube-api-access-p9jq4") pod "6d640be9-09ac-44ce-af77-68ea96aa6483" (UID: "6d640be9-09ac-44ce-af77-68ea96aa6483"). InnerVolumeSpecName "kube-api-access-p9jq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.630548 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9jq4\" (UniqueName: \"kubernetes.io/projected/6d640be9-09ac-44ce-af77-68ea96aa6483-kube-api-access-p9jq4\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.754161 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.834103 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc\") pod \"cb452d4d-d280-4035-9a47-d240d9051982\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.834172 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config\") pod \"cb452d4d-d280-4035-9a47-d240d9051982\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.834250 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2ssw\" (UniqueName: \"kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw\") pod \"cb452d4d-d280-4035-9a47-d240d9051982\" (UID: \"cb452d4d-d280-4035-9a47-d240d9051982\") " Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.837707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw" (OuterVolumeSpecName: "kube-api-access-v2ssw") pod "cb452d4d-d280-4035-9a47-d240d9051982" (UID: "cb452d4d-d280-4035-9a47-d240d9051982"). InnerVolumeSpecName "kube-api-access-v2ssw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.868924 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config" (OuterVolumeSpecName: "config") pod "cb452d4d-d280-4035-9a47-d240d9051982" (UID: "cb452d4d-d280-4035-9a47-d240d9051982"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.887738 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cb452d4d-d280-4035-9a47-d240d9051982" (UID: "cb452d4d-d280-4035-9a47-d240d9051982"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.936483 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.936508 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb452d4d-d280-4035-9a47-d240d9051982-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.936517 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2ssw\" (UniqueName: \"kubernetes.io/projected/cb452d4d-d280-4035-9a47-d240d9051982-kube-api-access-v2ssw\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.995630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2pszr" event={"ID":"6d640be9-09ac-44ce-af77-68ea96aa6483","Type":"ContainerDied","Data":"04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825"} Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.995671 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04eeb03a04bc54058aedb039ebc49fe2391d4a3495267556242cc0010f45c825" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.995671 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2pszr" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.997155 4959 generic.go:334] "Generic (PLEG): container finished" podID="cb452d4d-d280-4035-9a47-d240d9051982" containerID="991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f" exitCode=0 Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.997355 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.997604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" event={"ID":"cb452d4d-d280-4035-9a47-d240d9051982","Type":"ContainerDied","Data":"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f"} Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.997666 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4xmqn" event={"ID":"cb452d4d-d280-4035-9a47-d240d9051982","Type":"ContainerDied","Data":"29196dbd48b0a5216868e904688cae97384694480636df893c60588c6ebb11b1"} Oct 03 13:51:05 crc kubenswrapper[4959]: I1003 13:51:05.997714 4959 scope.go:117] "RemoveContainer" containerID="991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.018009 4959 scope.go:117] "RemoveContainer" containerID="ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.032355 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.043976 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4xmqn"] Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.057331 4959 scope.go:117] "RemoveContainer" containerID="991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f" Oct 03 13:51:06 crc kubenswrapper[4959]: E1003 13:51:06.057985 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f\": container with ID starting with 991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f not found: ID does not exist" containerID="991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.058014 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f"} err="failed to get container status \"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f\": rpc error: code = NotFound desc = could not find container \"991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f\": container with ID starting with 991d3da791c2293b0be3a58c00c31b805e7323e14426b59af2f1eca4c839563f not found: ID does not exist" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.058035 4959 scope.go:117] "RemoveContainer" containerID="ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2" Oct 03 13:51:06 crc kubenswrapper[4959]: E1003 13:51:06.058663 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2\": container with ID starting with ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2 not found: ID does not exist" containerID="ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.058730 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2"} err="failed to get container status \"ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2\": rpc error: code = NotFound desc = could not find container \"ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2\": container with ID starting with ade49065d994900ca355ab3799c643cd435ffde859726fc9b4874f524f1deed2 not found: ID does not exist" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.229608 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-949rt" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.345255 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms5dh\" (UniqueName: \"kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh\") pod \"56a3ee2c-e196-4d8d-a796-507963138289\" (UID: \"56a3ee2c-e196-4d8d-a796-507963138289\") " Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.348518 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh" (OuterVolumeSpecName: "kube-api-access-ms5dh") pod "56a3ee2c-e196-4d8d-a796-507963138289" (UID: "56a3ee2c-e196-4d8d-a796-507963138289"). InnerVolumeSpecName "kube-api-access-ms5dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:06 crc kubenswrapper[4959]: I1003 13:51:06.447751 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms5dh\" (UniqueName: \"kubernetes.io/projected/56a3ee2c-e196-4d8d-a796-507963138289-kube-api-access-ms5dh\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.007464 4959 generic.go:334] "Generic (PLEG): container finished" podID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerID="e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307" exitCode=0 Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.007542 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerDied","Data":"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307"} Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.010059 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-949rt" event={"ID":"56a3ee2c-e196-4d8d-a796-507963138289","Type":"ContainerDied","Data":"5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b"} Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.010084 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5295b9232cdf173183c0b7fcbbb3b06ff9d3deccea08dc955c0e6aafb17d0d1b" Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.010126 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-949rt" Oct 03 13:51:07 crc kubenswrapper[4959]: I1003 13:51:07.696241 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb452d4d-d280-4035-9a47-d240d9051982" path="/var/lib/kubelet/pods/cb452d4d-d280-4035-9a47-d240d9051982/volumes" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.027822 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerStarted","Data":"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6"} Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.028145 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.030491 4959 generic.go:334] "Generic (PLEG): container finished" podID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerID="1e12c25b84bbfc17840c8beb88edad3a093b4fd3bcbb1e87e14c1aaf4a2ed572" exitCode=0 Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.030884 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerDied","Data":"1e12c25b84bbfc17840c8beb88edad3a093b4fd3bcbb1e87e14c1aaf4a2ed572"} Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.034404 4959 generic.go:334] "Generic (PLEG): container finished" podID="9655f477-fe02-412d-b52d-6aea687e1e57" containerID="956042f00860508daaea28b26ecf1fd459cedf0fd52cb55d9ba130ca282f8dbc" exitCode=0 Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.034459 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-92xpj" event={"ID":"9655f477-fe02-412d-b52d-6aea687e1e57","Type":"ContainerDied","Data":"956042f00860508daaea28b26ecf1fd459cedf0fd52cb55d9ba130ca282f8dbc"} Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.066956 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.621090689 podStartE2EDuration="50.066936243s" podCreationTimestamp="2025-10-03 13:50:18 +0000 UTC" firstStartedPulling="2025-10-03 13:50:32.368046542 +0000 UTC m=+1201.571389959" lastFinishedPulling="2025-10-03 13:50:33.813892096 +0000 UTC m=+1203.017235513" observedRunningTime="2025-10-03 13:51:08.056378087 +0000 UTC m=+1237.259721524" watchObservedRunningTime="2025-10-03 13:51:08.066936243 +0000 UTC m=+1237.270279660" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.261815 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2jgb6"] Oct 03 13:51:08 crc kubenswrapper[4959]: E1003 13:51:08.262545 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d640be9-09ac-44ce-af77-68ea96aa6483" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262568 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d640be9-09ac-44ce-af77-68ea96aa6483" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: E1003 13:51:08.262591 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="init" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262598 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="init" Oct 03 13:51:08 crc kubenswrapper[4959]: E1003 13:51:08.262613 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a3ee2c-e196-4d8d-a796-507963138289" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262620 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a3ee2c-e196-4d8d-a796-507963138289" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: E1003 13:51:08.262630 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="dnsmasq-dns" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262636 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="dnsmasq-dns" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262780 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb452d4d-d280-4035-9a47-d240d9051982" containerName="dnsmasq-dns" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262793 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d640be9-09ac-44ce-af77-68ea96aa6483" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.262812 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a3ee2c-e196-4d8d-a796-507963138289" containerName="mariadb-database-create" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.263475 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.273788 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2jgb6"] Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.377526 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfhzp\" (UniqueName: \"kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp\") pod \"glance-db-create-2jgb6\" (UID: \"ce4b7381-3d0c-4067-a897-9b383b27a77c\") " pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.480039 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfhzp\" (UniqueName: \"kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp\") pod \"glance-db-create-2jgb6\" (UID: \"ce4b7381-3d0c-4067-a897-9b383b27a77c\") " pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.503314 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfhzp\" (UniqueName: \"kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp\") pod \"glance-db-create-2jgb6\" (UID: \"ce4b7381-3d0c-4067-a897-9b383b27a77c\") " pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.586295 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:08 crc kubenswrapper[4959]: I1003 13:51:08.769880 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.043029 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerStarted","Data":"73f4e64df58083370f19d3a84fd7e6d994163b876b512f98d2ee1e62e5754b19"} Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.071413 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.080926566 podStartE2EDuration="51.071384839s" podCreationTimestamp="2025-10-03 13:50:18 +0000 UTC" firstStartedPulling="2025-10-03 13:50:23.797398776 +0000 UTC m=+1193.000742193" lastFinishedPulling="2025-10-03 13:50:33.787857049 +0000 UTC m=+1202.991200466" observedRunningTime="2025-10-03 13:51:09.065840884 +0000 UTC m=+1238.269184301" watchObservedRunningTime="2025-10-03 13:51:09.071384839 +0000 UTC m=+1238.274728256" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.093719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2jgb6"] Oct 03 13:51:09 crc kubenswrapper[4959]: W1003 13:51:09.103252 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce4b7381_3d0c_4067_a897_9b383b27a77c.slice/crio-8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99 WatchSource:0}: Error finding container 8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99: Status 404 returned error can't find the container with id 8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99 Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.340769 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502571 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502638 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4cdj\" (UniqueName: \"kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502661 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502691 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502723 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.502787 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift\") pod \"9655f477-fe02-412d-b52d-6aea687e1e57\" (UID: \"9655f477-fe02-412d-b52d-6aea687e1e57\") " Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.503850 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.505728 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.515322 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.520966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj" (OuterVolumeSpecName: "kube-api-access-s4cdj") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "kube-api-access-s4cdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.527249 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.527392 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts" (OuterVolumeSpecName: "scripts") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.530711 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9655f477-fe02-412d-b52d-6aea687e1e57" (UID: "9655f477-fe02-412d-b52d-6aea687e1e57"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.604460 4959 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.604683 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4cdj\" (UniqueName: \"kubernetes.io/projected/9655f477-fe02-412d-b52d-6aea687e1e57-kube-api-access-s4cdj\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.604774 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.604836 4959 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.604897 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9655f477-fe02-412d-b52d-6aea687e1e57-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.605000 4959 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9655f477-fe02-412d-b52d-6aea687e1e57-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.605064 4959 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9655f477-fe02-412d-b52d-6aea687e1e57-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:09 crc kubenswrapper[4959]: I1003 13:51:09.818340 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.052845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-92xpj" event={"ID":"9655f477-fe02-412d-b52d-6aea687e1e57","Type":"ContainerDied","Data":"774fd1d79d4cadf2dfb5b1a7537f3510f284a3ce3107f456b21eb58a82e4fb52"} Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.052899 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="774fd1d79d4cadf2dfb5b1a7537f3510f284a3ce3107f456b21eb58a82e4fb52" Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.052869 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-92xpj" Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.055307 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce4b7381-3d0c-4067-a897-9b383b27a77c" containerID="79071188a4f44f92ce2ed170a1b1c0fd2bafcd83545facaa85075a5cda7b863a" exitCode=0 Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.055389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2jgb6" event={"ID":"ce4b7381-3d0c-4067-a897-9b383b27a77c","Type":"ContainerDied","Data":"79071188a4f44f92ce2ed170a1b1c0fd2bafcd83545facaa85075a5cda7b863a"} Oct 03 13:51:10 crc kubenswrapper[4959]: I1003 13:51:10.055442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2jgb6" event={"ID":"ce4b7381-3d0c-4067-a897-9b383b27a77c","Type":"ContainerStarted","Data":"8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99"} Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.370438 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.536915 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfhzp\" (UniqueName: \"kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp\") pod \"ce4b7381-3d0c-4067-a897-9b383b27a77c\" (UID: \"ce4b7381-3d0c-4067-a897-9b383b27a77c\") " Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.543342 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp" (OuterVolumeSpecName: "kube-api-access-wfhzp") pod "ce4b7381-3d0c-4067-a897-9b383b27a77c" (UID: "ce4b7381-3d0c-4067-a897-9b383b27a77c"). InnerVolumeSpecName "kube-api-access-wfhzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.639496 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfhzp\" (UniqueName: \"kubernetes.io/projected/ce4b7381-3d0c-4067-a897-9b383b27a77c-kube-api-access-wfhzp\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.946071 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.951356 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"swift-storage-0\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " pod="openstack/swift-storage-0" Oct 03 13:51:11 crc kubenswrapper[4959]: I1003 13:51:11.965050 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.072735 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2jgb6" event={"ID":"ce4b7381-3d0c-4067-a897-9b383b27a77c","Type":"ContainerDied","Data":"8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99"} Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.072770 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bdefdbe61e68e691c9a33bd747284a6ba766216602380aae7b52f74935aaa99" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.072822 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2jgb6" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.501644 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:51:12 crc kubenswrapper[4959]: W1003 13:51:12.505933 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29a02381_a818_4890_9c9a_089fc14e565d.slice/crio-fcb92ab976eceb4c5f0f226560479341a04e4c7eb7fe937996a882532e040c0d WatchSource:0}: Error finding container fcb92ab976eceb4c5f0f226560479341a04e4c7eb7fe937996a882532e040c0d: Status 404 returned error can't find the container with id fcb92ab976eceb4c5f0f226560479341a04e4c7eb7fe937996a882532e040c0d Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.735775 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7408-account-create-5dpjp"] Oct 03 13:51:12 crc kubenswrapper[4959]: E1003 13:51:12.736251 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9655f477-fe02-412d-b52d-6aea687e1e57" containerName="swift-ring-rebalance" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.736272 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9655f477-fe02-412d-b52d-6aea687e1e57" containerName="swift-ring-rebalance" Oct 03 13:51:12 crc kubenswrapper[4959]: E1003 13:51:12.736303 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4b7381-3d0c-4067-a897-9b383b27a77c" containerName="mariadb-database-create" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.736313 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4b7381-3d0c-4067-a897-9b383b27a77c" containerName="mariadb-database-create" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.736538 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4b7381-3d0c-4067-a897-9b383b27a77c" containerName="mariadb-database-create" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.736559 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9655f477-fe02-412d-b52d-6aea687e1e57" containerName="swift-ring-rebalance" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.737258 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.740048 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.744429 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7408-account-create-5dpjp"] Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.860660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfbl4\" (UniqueName: \"kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4\") pod \"keystone-7408-account-create-5dpjp\" (UID: \"363fcee0-62aa-4823-a5c8-c96c114bdeee\") " pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.962452 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfbl4\" (UniqueName: \"kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4\") pod \"keystone-7408-account-create-5dpjp\" (UID: \"363fcee0-62aa-4823-a5c8-c96c114bdeee\") " pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:12 crc kubenswrapper[4959]: I1003 13:51:12.982927 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfbl4\" (UniqueName: \"kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4\") pod \"keystone-7408-account-create-5dpjp\" (UID: \"363fcee0-62aa-4823-a5c8-c96c114bdeee\") " pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.053644 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.106505 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-aa19-account-create-j7rsg"] Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.108561 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.111474 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.134671 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-aa19-account-create-j7rsg"] Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.135343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"fcb92ab976eceb4c5f0f226560479341a04e4c7eb7fe937996a882532e040c0d"} Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.266432 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8z4n\" (UniqueName: \"kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n\") pod \"placement-aa19-account-create-j7rsg\" (UID: \"2d8b2056-b39e-4134-84b4-c911d90a4958\") " pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.365329 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" probeResult="failure" output=< Oct 03 13:51:13 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 13:51:13 crc kubenswrapper[4959]: > Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.367982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8z4n\" (UniqueName: \"kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n\") pod \"placement-aa19-account-create-j7rsg\" (UID: \"2d8b2056-b39e-4134-84b4-c911d90a4958\") " pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.387026 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8z4n\" (UniqueName: \"kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n\") pod \"placement-aa19-account-create-j7rsg\" (UID: \"2d8b2056-b39e-4134-84b4-c911d90a4958\") " pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.430829 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.495778 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:13 crc kubenswrapper[4959]: I1003 13:51:13.507374 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7408-account-create-5dpjp"] Oct 03 13:51:13 crc kubenswrapper[4959]: W1003 13:51:13.722864 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod363fcee0_62aa_4823_a5c8_c96c114bdeee.slice/crio-41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69 WatchSource:0}: Error finding container 41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69: Status 404 returned error can't find the container with id 41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69 Oct 03 13:51:14 crc kubenswrapper[4959]: I1003 13:51:14.143787 4959 generic.go:334] "Generic (PLEG): container finished" podID="363fcee0-62aa-4823-a5c8-c96c114bdeee" containerID="c2ef1d50c81020ba55c252cdc1ea93b3c7f0bf54d6a6ff2e7051d40381b130cf" exitCode=0 Oct 03 13:51:14 crc kubenswrapper[4959]: I1003 13:51:14.143861 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7408-account-create-5dpjp" event={"ID":"363fcee0-62aa-4823-a5c8-c96c114bdeee","Type":"ContainerDied","Data":"c2ef1d50c81020ba55c252cdc1ea93b3c7f0bf54d6a6ff2e7051d40381b130cf"} Oct 03 13:51:14 crc kubenswrapper[4959]: I1003 13:51:14.144218 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7408-account-create-5dpjp" event={"ID":"363fcee0-62aa-4823-a5c8-c96c114bdeee","Type":"ContainerStarted","Data":"41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69"} Oct 03 13:51:14 crc kubenswrapper[4959]: I1003 13:51:14.146115 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"0b5395c8c9b277726906ea76a40f04683a601fd733480edda15aba4179fba2b5"} Oct 03 13:51:14 crc kubenswrapper[4959]: I1003 13:51:14.254748 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-aa19-account-create-j7rsg"] Oct 03 13:51:14 crc kubenswrapper[4959]: W1003 13:51:14.257329 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d8b2056_b39e_4134_84b4_c911d90a4958.slice/crio-ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c WatchSource:0}: Error finding container ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c: Status 404 returned error can't find the container with id ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.157708 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"bed191a2d6ad3263372ef130bd3b54ebdd2a1a844aa9131fc1eac84370b2993e"} Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.158092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"34ae11a48814fa28d689447bfba27eedc369d31d57a2d21cecd65f5c3ab9be8a"} Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.158115 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"fc9e2bfb000766ae8566eb0974e5a4ea4297a6971e4c71891ab0065f78e71574"} Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.161756 4959 generic.go:334] "Generic (PLEG): container finished" podID="2d8b2056-b39e-4134-84b4-c911d90a4958" containerID="3dc40507c120959d206031c3611237f52e593b9449af8cabe30688abc704507f" exitCode=0 Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.161849 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa19-account-create-j7rsg" event={"ID":"2d8b2056-b39e-4134-84b4-c911d90a4958","Type":"ContainerDied","Data":"3dc40507c120959d206031c3611237f52e593b9449af8cabe30688abc704507f"} Oct 03 13:51:15 crc kubenswrapper[4959]: I1003 13:51:15.161890 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa19-account-create-j7rsg" event={"ID":"2d8b2056-b39e-4134-84b4-c911d90a4958","Type":"ContainerStarted","Data":"ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c"} Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.554858 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.561878 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.661109 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfbl4\" (UniqueName: \"kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4\") pod \"363fcee0-62aa-4823-a5c8-c96c114bdeee\" (UID: \"363fcee0-62aa-4823-a5c8-c96c114bdeee\") " Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.661247 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8z4n\" (UniqueName: \"kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n\") pod \"2d8b2056-b39e-4134-84b4-c911d90a4958\" (UID: \"2d8b2056-b39e-4134-84b4-c911d90a4958\") " Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.674890 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4" (OuterVolumeSpecName: "kube-api-access-mfbl4") pod "363fcee0-62aa-4823-a5c8-c96c114bdeee" (UID: "363fcee0-62aa-4823-a5c8-c96c114bdeee"). InnerVolumeSpecName "kube-api-access-mfbl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.674974 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n" (OuterVolumeSpecName: "kube-api-access-l8z4n") pod "2d8b2056-b39e-4134-84b4-c911d90a4958" (UID: "2d8b2056-b39e-4134-84b4-c911d90a4958"). InnerVolumeSpecName "kube-api-access-l8z4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.763606 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfbl4\" (UniqueName: \"kubernetes.io/projected/363fcee0-62aa-4823-a5c8-c96c114bdeee-kube-api-access-mfbl4\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:16 crc kubenswrapper[4959]: I1003 13:51:16.763655 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8z4n\" (UniqueName: \"kubernetes.io/projected/2d8b2056-b39e-4134-84b4-c911d90a4958-kube-api-access-l8z4n\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.182494 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"51c570be58868ca7518ace8ff746ead4f09d193a7fef31832eeba7dc57147fd9"} Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.182542 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"e40663af33c9b4d1c19be468f7912534915bb3fba53c5989f5776aa52e428838"} Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.184943 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7408-account-create-5dpjp" Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.184979 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7408-account-create-5dpjp" event={"ID":"363fcee0-62aa-4823-a5c8-c96c114bdeee","Type":"ContainerDied","Data":"41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69"} Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.185007 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41fd5ed7ec7f2268ebe294e93f1577f49916b07212912a5bc37d7c30edb9ee69" Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.186374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-aa19-account-create-j7rsg" event={"ID":"2d8b2056-b39e-4134-84b4-c911d90a4958","Type":"ContainerDied","Data":"ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c"} Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.186509 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff670fc87ca16b28d88d3037a5bb1703244d36da9c99a8ca2739325741cacc6c" Oct 03 13:51:17 crc kubenswrapper[4959]: I1003 13:51:17.186391 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-aa19-account-create-j7rsg" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.207686 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"326cf0aa34a60cab6dd4c8c253341c34b40b93848f039d7eab65897e840f352f"} Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.207972 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"27d21e8e01afbe1d304e1134184443c9d6a5ca00c5cf331aa5561d1989f0cb19"} Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.305713 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6565-account-create-8k2w6"] Oct 03 13:51:18 crc kubenswrapper[4959]: E1003 13:51:18.306056 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363fcee0-62aa-4823-a5c8-c96c114bdeee" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.306077 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="363fcee0-62aa-4823-a5c8-c96c114bdeee" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: E1003 13:51:18.306095 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8b2056-b39e-4134-84b4-c911d90a4958" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.306102 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8b2056-b39e-4134-84b4-c911d90a4958" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.306263 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="363fcee0-62aa-4823-a5c8-c96c114bdeee" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.306296 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8b2056-b39e-4134-84b4-c911d90a4958" containerName="mariadb-account-create" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.306872 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.310671 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.314827 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6565-account-create-8k2w6"] Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.366720 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" probeResult="failure" output=< Oct 03 13:51:18 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 13:51:18 crc kubenswrapper[4959]: > Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.403264 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74k2n\" (UniqueName: \"kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n\") pod \"glance-6565-account-create-8k2w6\" (UID: \"43dd7882-778f-40c0-b7d4-757f090605d3\") " pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.421785 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.504548 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74k2n\" (UniqueName: \"kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n\") pod \"glance-6565-account-create-8k2w6\" (UID: \"43dd7882-778f-40c0-b7d4-757f090605d3\") " pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.523982 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74k2n\" (UniqueName: \"kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n\") pod \"glance-6565-account-create-8k2w6\" (UID: \"43dd7882-778f-40c0-b7d4-757f090605d3\") " pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.611530 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k4vvh-config-62ntq"] Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.614257 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.621832 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4vvh-config-62ntq"] Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.626220 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.626487 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.708355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cxtz\" (UniqueName: \"kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.710033 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.710313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.710459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.710530 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.710586 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812467 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812501 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812524 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cxtz\" (UniqueName: \"kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812578 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812884 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.812971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.816001 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.816597 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.834925 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cxtz\" (UniqueName: \"kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz\") pod \"ovn-controller-k4vvh-config-62ntq\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:18 crc kubenswrapper[4959]: I1003 13:51:18.929754 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:19 crc kubenswrapper[4959]: I1003 13:51:19.220020 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"23005919d7ab8893cd7eced93743eb0b32520fcd64407c9e2193608b3091452f"} Oct 03 13:51:19 crc kubenswrapper[4959]: I1003 13:51:19.344564 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6565-account-create-8k2w6"] Oct 03 13:51:19 crc kubenswrapper[4959]: W1003 13:51:19.348493 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43dd7882_778f_40c0_b7d4_757f090605d3.slice/crio-cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882 WatchSource:0}: Error finding container cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882: Status 404 returned error can't find the container with id cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882 Oct 03 13:51:19 crc kubenswrapper[4959]: I1003 13:51:19.426231 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4vvh-config-62ntq"] Oct 03 13:51:19 crc kubenswrapper[4959]: W1003 13:51:19.429287 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0b21d38_ab6c_4725_967a_6c0ff3161785.slice/crio-fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772 WatchSource:0}: Error finding container fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772: Status 404 returned error can't find the container with id fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772 Oct 03 13:51:19 crc kubenswrapper[4959]: I1003 13:51:19.821379 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.203397 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.204018 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-mksg4"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.205302 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.213752 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mksg4"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.240822 4959 generic.go:334] "Generic (PLEG): container finished" podID="43dd7882-778f-40c0-b7d4-757f090605d3" containerID="fa4db0133a9fd539ffb0778a58527848be291d332cab66fa55c6d83908ac5d53" exitCode=0 Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.240880 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6565-account-create-8k2w6" event={"ID":"43dd7882-778f-40c0-b7d4-757f090605d3","Type":"ContainerDied","Data":"fa4db0133a9fd539ffb0778a58527848be291d332cab66fa55c6d83908ac5d53"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.240920 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6565-account-create-8k2w6" event={"ID":"43dd7882-778f-40c0-b7d4-757f090605d3","Type":"ContainerStarted","Data":"cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.248672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh-config-62ntq" event={"ID":"f0b21d38-ab6c-4725-967a-6c0ff3161785","Type":"ContainerStarted","Data":"87357d5f6f8004473f56f3dd117b5d6a1b35ed9b5d28836c98184b8c60d36601"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.248975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh-config-62ntq" event={"ID":"f0b21d38-ab6c-4725-967a-6c0ff3161785","Type":"ContainerStarted","Data":"fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.267306 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7pxk\" (UniqueName: \"kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk\") pod \"barbican-db-create-mksg4\" (UID: \"d93957ba-a96d-4a52-ab5c-46c80c04f451\") " pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.287939 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"8d597ede10216948eca70fee9a0011ee11ae068e5bb066f3921145f9f96049e6"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.287984 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"004034aa9b447a038bca1e3cffcf1b4fa3f5c22317a682340c9b6936b0225942"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.287993 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"e93efd61469a9f4593c4dda8ba711b8d800d8760fc00d01dc03296a37a56265e"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.288001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"30ca3bfd6f09224ce61168200372b479678d8278ab9c2dd7cba0dd99816f2b87"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.288009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"a3d3d328227c680fc55d1e2c7cd597cae86b320ddc73849d22692bc5ba674e45"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.288018 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerStarted","Data":"6000516424e24bea6505665e93fef7a122ed78ffe7c2ba42d237e867092bc2dd"} Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.315766 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6nt5n"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.316709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.349601 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-k4vvh-config-62ntq" podStartSLOduration=2.349583866 podStartE2EDuration="2.349583866s" podCreationTimestamp="2025-10-03 13:51:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:20.337427781 +0000 UTC m=+1249.540771208" watchObservedRunningTime="2025-10-03 13:51:20.349583866 +0000 UTC m=+1249.552927283" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.355696 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6nt5n"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.377516 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54x7z\" (UniqueName: \"kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z\") pod \"cinder-db-create-6nt5n\" (UID: \"2f0ebab1-59f6-4904-b830-b252e06f03e9\") " pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.377673 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7pxk\" (UniqueName: \"kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk\") pod \"barbican-db-create-mksg4\" (UID: \"d93957ba-a96d-4a52-ab5c-46c80c04f451\") " pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.394446 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.914007325 podStartE2EDuration="26.394422654s" podCreationTimestamp="2025-10-03 13:50:54 +0000 UTC" firstStartedPulling="2025-10-03 13:51:12.509124382 +0000 UTC m=+1241.712467799" lastFinishedPulling="2025-10-03 13:51:18.989539721 +0000 UTC m=+1248.192883128" observedRunningTime="2025-10-03 13:51:20.3905487 +0000 UTC m=+1249.593892127" watchObservedRunningTime="2025-10-03 13:51:20.394422654 +0000 UTC m=+1249.597766071" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.431440 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7pxk\" (UniqueName: \"kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk\") pod \"barbican-db-create-mksg4\" (UID: \"d93957ba-a96d-4a52-ab5c-46c80c04f451\") " pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.479110 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54x7z\" (UniqueName: \"kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z\") pod \"cinder-db-create-6nt5n\" (UID: \"2f0ebab1-59f6-4904-b830-b252e06f03e9\") " pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.505565 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54x7z\" (UniqueName: \"kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z\") pod \"cinder-db-create-6nt5n\" (UID: \"2f0ebab1-59f6-4904-b830-b252e06f03e9\") " pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.522770 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.523302 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-dxntm"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.524586 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.581905 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dxntm"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.657619 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.671072 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xbd64"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.672351 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.676585 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.676789 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.676824 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.677026 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8gbd" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.681719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xbd64"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.684099 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkz5q\" (UniqueName: \"kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q\") pod \"neutron-db-create-dxntm\" (UID: \"6e634f23-0fe8-438b-9657-d719b3f01205\") " pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.788942 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.788999 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s9qq\" (UniqueName: \"kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.789021 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkz5q\" (UniqueName: \"kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q\") pod \"neutron-db-create-dxntm\" (UID: \"6e634f23-0fe8-438b-9657-d719b3f01205\") " pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.789097 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.822950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkz5q\" (UniqueName: \"kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q\") pod \"neutron-db-create-dxntm\" (UID: \"6e634f23-0fe8-438b-9657-d719b3f01205\") " pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.879478 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.880993 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.890484 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.890593 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.890682 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.890929 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.890984 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.891053 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njprn\" (UniqueName: \"kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.891079 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s9qq\" (UniqueName: \"kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.891098 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.891168 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.891388 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.898136 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.903562 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.925917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s9qq\" (UniqueName: \"kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq\") pod \"keystone-db-sync-xbd64\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.926581 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992601 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njprn\" (UniqueName: \"kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992763 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992798 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.992828 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.993612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.994393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.994940 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.995276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:20 crc kubenswrapper[4959]: I1003 13:51:20.996504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.020798 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njprn\" (UniqueName: \"kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn\") pod \"dnsmasq-dns-77585f5f8c-jzpz5\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.055650 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.081131 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.205721 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mksg4"] Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.212140 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:21 crc kubenswrapper[4959]: W1003 13:51:21.219255 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd93957ba_a96d_4a52_ab5c_46c80c04f451.slice/crio-6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b WatchSource:0}: Error finding container 6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b: Status 404 returned error can't find the container with id 6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.301622 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mksg4" event={"ID":"d93957ba-a96d-4a52-ab5c-46c80c04f451","Type":"ContainerStarted","Data":"6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b"} Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.304057 4959 generic.go:334] "Generic (PLEG): container finished" podID="f0b21d38-ab6c-4725-967a-6c0ff3161785" containerID="87357d5f6f8004473f56f3dd117b5d6a1b35ed9b5d28836c98184b8c60d36601" exitCode=0 Oct 03 13:51:21 crc kubenswrapper[4959]: I1003 13:51:21.304234 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh-config-62ntq" event={"ID":"f0b21d38-ab6c-4725-967a-6c0ff3161785","Type":"ContainerDied","Data":"87357d5f6f8004473f56f3dd117b5d6a1b35ed9b5d28836c98184b8c60d36601"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:21.366183 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6nt5n"] Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:21.406966 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dxntm"] Oct 03 13:51:22 crc kubenswrapper[4959]: W1003 13:51:21.424759 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e634f23_0fe8_438b_9657_d719b3f01205.slice/crio-aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a WatchSource:0}: Error finding container aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a: Status 404 returned error can't find the container with id aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:21.533629 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xbd64"] Oct 03 13:51:22 crc kubenswrapper[4959]: W1003 13:51:21.542229 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb4495d5_4d1f_4ec5_af52_13a6a444148a.slice/crio-292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324 WatchSource:0}: Error finding container 292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324: Status 404 returned error can't find the container with id 292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324 Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.313328 4959 generic.go:334] "Generic (PLEG): container finished" podID="6e634f23-0fe8-438b-9657-d719b3f01205" containerID="64c14a000a4fd84bbb5d50a1259b64583f506161d19b324050350a8e3f0d3acf" exitCode=0 Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.313475 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dxntm" event={"ID":"6e634f23-0fe8-438b-9657-d719b3f01205","Type":"ContainerDied","Data":"64c14a000a4fd84bbb5d50a1259b64583f506161d19b324050350a8e3f0d3acf"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.313676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dxntm" event={"ID":"6e634f23-0fe8-438b-9657-d719b3f01205","Type":"ContainerStarted","Data":"aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.319293 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f0ebab1-59f6-4904-b830-b252e06f03e9" containerID="05c8d7b052abc4c1170e3e6ba142cf474e73af12bf8462a593d2a4d55762f344" exitCode=0 Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.319366 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6nt5n" event={"ID":"2f0ebab1-59f6-4904-b830-b252e06f03e9","Type":"ContainerDied","Data":"05c8d7b052abc4c1170e3e6ba142cf474e73af12bf8462a593d2a4d55762f344"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.319393 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6nt5n" event={"ID":"2f0ebab1-59f6-4904-b830-b252e06f03e9","Type":"ContainerStarted","Data":"bdb16ff64abedd4342e37f5082f67ff1b3a185d2ec5cb0a5d4d70aea9bac48fe"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.322787 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6565-account-create-8k2w6" event={"ID":"43dd7882-778f-40c0-b7d4-757f090605d3","Type":"ContainerDied","Data":"cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.322921 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdd3929b52f2bae8a48faf09bc65d569a4d8d8a81a1df153a5716a1b7a779882" Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.327674 4959 generic.go:334] "Generic (PLEG): container finished" podID="d93957ba-a96d-4a52-ab5c-46c80c04f451" containerID="6f930a2e79f58fb03314be9aa1247f4c2514347820d5bc31d5b2e66c097d5a69" exitCode=0 Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.327815 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mksg4" event={"ID":"d93957ba-a96d-4a52-ab5c-46c80c04f451","Type":"ContainerDied","Data":"6f930a2e79f58fb03314be9aa1247f4c2514347820d5bc31d5b2e66c097d5a69"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.370763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbd64" event={"ID":"db4495d5-4d1f-4ec5-af52-13a6a444148a","Type":"ContainerStarted","Data":"292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324"} Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.395802 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.446668 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74k2n\" (UniqueName: \"kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n\") pod \"43dd7882-778f-40c0-b7d4-757f090605d3\" (UID: \"43dd7882-778f-40c0-b7d4-757f090605d3\") " Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.455679 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n" (OuterVolumeSpecName: "kube-api-access-74k2n") pod "43dd7882-778f-40c0-b7d4-757f090605d3" (UID: "43dd7882-778f-40c0-b7d4-757f090605d3"). InnerVolumeSpecName "kube-api-access-74k2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.489505 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.548142 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74k2n\" (UniqueName: \"kubernetes.io/projected/43dd7882-778f-40c0-b7d4-757f090605d3-kube-api-access-74k2n\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:22 crc kubenswrapper[4959]: I1003 13:51:22.856256 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.055723 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.055818 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cxtz\" (UniqueName: \"kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.055880 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.055911 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.055960 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.056053 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts\") pod \"f0b21d38-ab6c-4725-967a-6c0ff3161785\" (UID: \"f0b21d38-ab6c-4725-967a-6c0ff3161785\") " Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.057236 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.058096 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts" (OuterVolumeSpecName: "scripts") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.058705 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.058740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.058739 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run" (OuterVolumeSpecName: "var-run") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.063215 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz" (OuterVolumeSpecName: "kube-api-access-6cxtz") pod "f0b21d38-ab6c-4725-967a-6c0ff3161785" (UID: "f0b21d38-ab6c-4725-967a-6c0ff3161785"). InnerVolumeSpecName "kube-api-access-6cxtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157592 4959 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157625 4959 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157635 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0b21d38-ab6c-4725-967a-6c0ff3161785-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157644 4959 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157654 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0b21d38-ab6c-4725-967a-6c0ff3161785-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.157662 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cxtz\" (UniqueName: \"kubernetes.io/projected/f0b21d38-ab6c-4725-967a-6c0ff3161785-kube-api-access-6cxtz\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.365251 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-k4vvh" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.386070 4959 generic.go:334] "Generic (PLEG): container finished" podID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerID="1d1fedaf479efcb97775d0f1d8772354aad002fe93020e429df0b895fb0d9726" exitCode=0 Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.386165 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" event={"ID":"bcee75a9-6e74-4e2e-b196-623472e8b6ca","Type":"ContainerDied","Data":"1d1fedaf479efcb97775d0f1d8772354aad002fe93020e429df0b895fb0d9726"} Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.386204 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" event={"ID":"bcee75a9-6e74-4e2e-b196-623472e8b6ca","Type":"ContainerStarted","Data":"c7320710702f8bbb405dfbc439b1a9a26d3af1d04811be66eebcd6d6b80d87c6"} Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.390931 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6565-account-create-8k2w6" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.391483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh-config-62ntq" event={"ID":"f0b21d38-ab6c-4725-967a-6c0ff3161785","Type":"ContainerDied","Data":"fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772"} Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.391648 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe385c4c1520b90e5c1a9001b5b86d1422a616557b17f968d3743c2b50b37772" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.391835 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh-config-62ntq" Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.487627 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k4vvh-config-62ntq"] Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.504138 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-k4vvh-config-62ntq"] Oct 03 13:51:23 crc kubenswrapper[4959]: I1003 13:51:23.703664 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b21d38-ab6c-4725-967a-6c0ff3161785" path="/var/lib/kubelet/pods/f0b21d38-ab6c-4725-967a-6c0ff3161785/volumes" Oct 03 13:51:24 crc kubenswrapper[4959]: I1003 13:51:24.409372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" event={"ID":"bcee75a9-6e74-4e2e-b196-623472e8b6ca","Type":"ContainerStarted","Data":"4791f6cd327460a52e47de9102371966ebd3798e713c07095beacd598e7d3450"} Oct 03 13:51:24 crc kubenswrapper[4959]: I1003 13:51:24.409567 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:24 crc kubenswrapper[4959]: I1003 13:51:24.433098 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" podStartSLOduration=4.43307794 podStartE2EDuration="4.43307794s" podCreationTimestamp="2025-10-03 13:51:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:24.432470775 +0000 UTC m=+1253.635814232" watchObservedRunningTime="2025-10-03 13:51:24.43307794 +0000 UTC m=+1253.636421357" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.427151 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dxntm" event={"ID":"6e634f23-0fe8-438b-9657-d719b3f01205","Type":"ContainerDied","Data":"aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a"} Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.427646 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aed459e7bf455c38f72bfb174f6d1c4abc9fb803b7218fd11c974cfa9597978a" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.428855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6nt5n" event={"ID":"2f0ebab1-59f6-4904-b830-b252e06f03e9","Type":"ContainerDied","Data":"bdb16ff64abedd4342e37f5082f67ff1b3a185d2ec5cb0a5d4d70aea9bac48fe"} Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.428893 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdb16ff64abedd4342e37f5082f67ff1b3a185d2ec5cb0a5d4d70aea9bac48fe" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.430658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mksg4" event={"ID":"d93957ba-a96d-4a52-ab5c-46c80c04f451","Type":"ContainerDied","Data":"6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b"} Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.430711 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6748c957a929225a7b3d9892cf9e3849e7173a116eb54026a3484b64277c236b" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.521169 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.563558 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.569267 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.672422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54x7z\" (UniqueName: \"kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z\") pod \"2f0ebab1-59f6-4904-b830-b252e06f03e9\" (UID: \"2f0ebab1-59f6-4904-b830-b252e06f03e9\") " Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.672610 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkz5q\" (UniqueName: \"kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q\") pod \"6e634f23-0fe8-438b-9657-d719b3f01205\" (UID: \"6e634f23-0fe8-438b-9657-d719b3f01205\") " Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.672684 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7pxk\" (UniqueName: \"kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk\") pod \"d93957ba-a96d-4a52-ab5c-46c80c04f451\" (UID: \"d93957ba-a96d-4a52-ab5c-46c80c04f451\") " Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.676641 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q" (OuterVolumeSpecName: "kube-api-access-nkz5q") pod "6e634f23-0fe8-438b-9657-d719b3f01205" (UID: "6e634f23-0fe8-438b-9657-d719b3f01205"). InnerVolumeSpecName "kube-api-access-nkz5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.677940 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z" (OuterVolumeSpecName: "kube-api-access-54x7z") pod "2f0ebab1-59f6-4904-b830-b252e06f03e9" (UID: "2f0ebab1-59f6-4904-b830-b252e06f03e9"). InnerVolumeSpecName "kube-api-access-54x7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.678052 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk" (OuterVolumeSpecName: "kube-api-access-d7pxk") pod "d93957ba-a96d-4a52-ab5c-46c80c04f451" (UID: "d93957ba-a96d-4a52-ab5c-46c80c04f451"). InnerVolumeSpecName "kube-api-access-d7pxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.774612 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkz5q\" (UniqueName: \"kubernetes.io/projected/6e634f23-0fe8-438b-9657-d719b3f01205-kube-api-access-nkz5q\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.774652 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7pxk\" (UniqueName: \"kubernetes.io/projected/d93957ba-a96d-4a52-ab5c-46c80c04f451-kube-api-access-d7pxk\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:26 crc kubenswrapper[4959]: I1003 13:51:26.774667 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54x7z\" (UniqueName: \"kubernetes.io/projected/2f0ebab1-59f6-4904-b830-b252e06f03e9-kube-api-access-54x7z\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:27 crc kubenswrapper[4959]: I1003 13:51:27.440089 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mksg4" Oct 03 13:51:27 crc kubenswrapper[4959]: I1003 13:51:27.440095 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbd64" event={"ID":"db4495d5-4d1f-4ec5-af52-13a6a444148a","Type":"ContainerStarted","Data":"a7bafb2fc033fe6372313d85d5173772ae3d7f9d3ac72d25597f8bd36ea8a953"} Oct 03 13:51:27 crc kubenswrapper[4959]: I1003 13:51:27.440149 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dxntm" Oct 03 13:51:27 crc kubenswrapper[4959]: I1003 13:51:27.440149 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6nt5n" Oct 03 13:51:27 crc kubenswrapper[4959]: I1003 13:51:27.468042 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xbd64" podStartSLOduration=2.694375007 podStartE2EDuration="7.468019009s" podCreationTimestamp="2025-10-03 13:51:20 +0000 UTC" firstStartedPulling="2025-10-03 13:51:21.544472552 +0000 UTC m=+1250.747815969" lastFinishedPulling="2025-10-03 13:51:26.318116554 +0000 UTC m=+1255.521459971" observedRunningTime="2025-10-03 13:51:27.465459037 +0000 UTC m=+1256.668802464" watchObservedRunningTime="2025-10-03 13:51:27.468019009 +0000 UTC m=+1256.671362436" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440172 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-cm7bx"] Oct 03 13:51:28 crc kubenswrapper[4959]: E1003 13:51:28.440585 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b21d38-ab6c-4725-967a-6c0ff3161785" containerName="ovn-config" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440604 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b21d38-ab6c-4725-967a-6c0ff3161785" containerName="ovn-config" Oct 03 13:51:28 crc kubenswrapper[4959]: E1003 13:51:28.440626 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93957ba-a96d-4a52-ab5c-46c80c04f451" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440635 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93957ba-a96d-4a52-ab5c-46c80c04f451" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: E1003 13:51:28.440660 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0ebab1-59f6-4904-b830-b252e06f03e9" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440669 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0ebab1-59f6-4904-b830-b252e06f03e9" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: E1003 13:51:28.440686 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43dd7882-778f-40c0-b7d4-757f090605d3" containerName="mariadb-account-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440694 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="43dd7882-778f-40c0-b7d4-757f090605d3" containerName="mariadb-account-create" Oct 03 13:51:28 crc kubenswrapper[4959]: E1003 13:51:28.440726 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e634f23-0fe8-438b-9657-d719b3f01205" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440734 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e634f23-0fe8-438b-9657-d719b3f01205" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440957 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93957ba-a96d-4a52-ab5c-46c80c04f451" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440973 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0ebab1-59f6-4904-b830-b252e06f03e9" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.440994 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e634f23-0fe8-438b-9657-d719b3f01205" containerName="mariadb-database-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.441010 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="43dd7882-778f-40c0-b7d4-757f090605d3" containerName="mariadb-account-create" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.441026 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b21d38-ab6c-4725-967a-6c0ff3161785" containerName="ovn-config" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.441711 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.444133 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-blm84" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.444377 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.454250 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-cm7bx"] Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.599486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr56m\" (UniqueName: \"kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.599720 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.599827 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.600170 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.702372 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.702572 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr56m\" (UniqueName: \"kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.702664 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.702724 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.709353 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.709567 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.712631 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.723841 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr56m\" (UniqueName: \"kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m\") pod \"glance-db-sync-cm7bx\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:28 crc kubenswrapper[4959]: I1003 13:51:28.764568 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cm7bx" Oct 03 13:51:29 crc kubenswrapper[4959]: I1003 13:51:29.288678 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-cm7bx"] Oct 03 13:51:29 crc kubenswrapper[4959]: W1003 13:51:29.296320 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc41cfce6_87b2_4ff2_8697_e739f3fa1ed7.slice/crio-e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811 WatchSource:0}: Error finding container e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811: Status 404 returned error can't find the container with id e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811 Oct 03 13:51:29 crc kubenswrapper[4959]: I1003 13:51:29.475373 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cm7bx" event={"ID":"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7","Type":"ContainerStarted","Data":"e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811"} Oct 03 13:51:30 crc kubenswrapper[4959]: I1003 13:51:30.484380 4959 generic.go:334] "Generic (PLEG): container finished" podID="db4495d5-4d1f-4ec5-af52-13a6a444148a" containerID="a7bafb2fc033fe6372313d85d5173772ae3d7f9d3ac72d25597f8bd36ea8a953" exitCode=0 Oct 03 13:51:30 crc kubenswrapper[4959]: I1003 13:51:30.484456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbd64" event={"ID":"db4495d5-4d1f-4ec5-af52-13a6a444148a","Type":"ContainerDied","Data":"a7bafb2fc033fe6372313d85d5173772ae3d7f9d3ac72d25597f8bd36ea8a953"} Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.214467 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.285552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.285823 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-7lrmq" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="dnsmasq-dns" containerID="cri-o://7bd9b89a81b4425bd0f9e9824cc5249a65423d9d48ba6caf70db91ea28349056" gracePeriod=10 Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.520262 4959 generic.go:334] "Generic (PLEG): container finished" podID="593147fe-5977-46ff-8f90-627e3a48142c" containerID="7bd9b89a81b4425bd0f9e9824cc5249a65423d9d48ba6caf70db91ea28349056" exitCode=0 Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.520483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7lrmq" event={"ID":"593147fe-5977-46ff-8f90-627e3a48142c","Type":"ContainerDied","Data":"7bd9b89a81b4425bd0f9e9824cc5249a65423d9d48ba6caf70db91ea28349056"} Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.882594 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.898124 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965789 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data\") pod \"db4495d5-4d1f-4ec5-af52-13a6a444148a\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965831 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s9qq\" (UniqueName: \"kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq\") pod \"db4495d5-4d1f-4ec5-af52-13a6a444148a\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965871 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb\") pod \"593147fe-5977-46ff-8f90-627e3a48142c\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config\") pod \"593147fe-5977-46ff-8f90-627e3a48142c\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb\") pod \"593147fe-5977-46ff-8f90-627e3a48142c\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.965998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle\") pod \"db4495d5-4d1f-4ec5-af52-13a6a444148a\" (UID: \"db4495d5-4d1f-4ec5-af52-13a6a444148a\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.966028 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc\") pod \"593147fe-5977-46ff-8f90-627e3a48142c\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.966065 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pq7c\" (UniqueName: \"kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c\") pod \"593147fe-5977-46ff-8f90-627e3a48142c\" (UID: \"593147fe-5977-46ff-8f90-627e3a48142c\") " Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.971330 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c" (OuterVolumeSpecName: "kube-api-access-8pq7c") pod "593147fe-5977-46ff-8f90-627e3a48142c" (UID: "593147fe-5977-46ff-8f90-627e3a48142c"). InnerVolumeSpecName "kube-api-access-8pq7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.972845 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq" (OuterVolumeSpecName: "kube-api-access-6s9qq") pod "db4495d5-4d1f-4ec5-af52-13a6a444148a" (UID: "db4495d5-4d1f-4ec5-af52-13a6a444148a"). InnerVolumeSpecName "kube-api-access-6s9qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:31 crc kubenswrapper[4959]: I1003 13:51:31.997625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db4495d5-4d1f-4ec5-af52-13a6a444148a" (UID: "db4495d5-4d1f-4ec5-af52-13a6a444148a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.012977 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "593147fe-5977-46ff-8f90-627e3a48142c" (UID: "593147fe-5977-46ff-8f90-627e3a48142c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.017618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "593147fe-5977-46ff-8f90-627e3a48142c" (UID: "593147fe-5977-46ff-8f90-627e3a48142c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.024054 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "593147fe-5977-46ff-8f90-627e3a48142c" (UID: "593147fe-5977-46ff-8f90-627e3a48142c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.027001 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data" (OuterVolumeSpecName: "config-data") pod "db4495d5-4d1f-4ec5-af52-13a6a444148a" (UID: "db4495d5-4d1f-4ec5-af52-13a6a444148a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.033798 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config" (OuterVolumeSpecName: "config") pod "593147fe-5977-46ff-8f90-627e3a48142c" (UID: "593147fe-5977-46ff-8f90-627e3a48142c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071365 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071403 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071414 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071423 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071432 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pq7c\" (UniqueName: \"kubernetes.io/projected/593147fe-5977-46ff-8f90-627e3a48142c-kube-api-access-8pq7c\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071440 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4495d5-4d1f-4ec5-af52-13a6a444148a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071449 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s9qq\" (UniqueName: \"kubernetes.io/projected/db4495d5-4d1f-4ec5-af52-13a6a444148a-kube-api-access-6s9qq\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.071457 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/593147fe-5977-46ff-8f90-627e3a48142c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.548681 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbd64" event={"ID":"db4495d5-4d1f-4ec5-af52-13a6a444148a","Type":"ContainerDied","Data":"292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324"} Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.548733 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="292e3cb0f4bf640b1bf876a69666f153cba370d384f73f2733ef8944a407a324" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.548742 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbd64" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.557247 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7lrmq" event={"ID":"593147fe-5977-46ff-8f90-627e3a48142c","Type":"ContainerDied","Data":"483a376d93b8ec0091ddaaf53035128b8f2ef19e92e59348ac63a5c1a313f3e4"} Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.557305 4959 scope.go:117] "RemoveContainer" containerID="7bd9b89a81b4425bd0f9e9824cc5249a65423d9d48ba6caf70db91ea28349056" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.557479 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7lrmq" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.617096 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.617324 4959 scope.go:117] "RemoveContainer" containerID="13e2b340838fa505ce2c1b982d466277a68ebe4aaad9e667134be02bc57d0a38" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.626435 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7lrmq"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.684659 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-l26lm"] Oct 03 13:51:32 crc kubenswrapper[4959]: E1003 13:51:32.685147 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="dnsmasq-dns" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.685174 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="dnsmasq-dns" Oct 03 13:51:32 crc kubenswrapper[4959]: E1003 13:51:32.685232 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="init" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.685242 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="init" Oct 03 13:51:32 crc kubenswrapper[4959]: E1003 13:51:32.685256 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4495d5-4d1f-4ec5-af52-13a6a444148a" containerName="keystone-db-sync" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.685267 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4495d5-4d1f-4ec5-af52-13a6a444148a" containerName="keystone-db-sync" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.685491 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="593147fe-5977-46ff-8f90-627e3a48142c" containerName="dnsmasq-dns" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.685523 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4495d5-4d1f-4ec5-af52-13a6a444148a" containerName="keystone-db-sync" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.686774 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.698448 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-gczs6"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.703157 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.709414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8gbd" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.709746 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.709941 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.710097 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.714247 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-l26lm"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.726983 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gczs6"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.853945 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.860379 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.867014 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.867447 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.874792 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887250 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887293 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnpvb\" (UniqueName: \"kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887379 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887415 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887452 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887485 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887539 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887559 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887590 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887629 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887690 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdfxv\" (UniqueName: \"kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887723 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887797 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887879 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887955 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4m47\" (UniqueName: \"kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.887989 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.888018 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.888063 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.989894 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdfxv\" (UniqueName: \"kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.989938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.989976 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4m47\" (UniqueName: \"kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990062 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990084 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990105 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990147 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990164 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnpvb\" (UniqueName: \"kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990261 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990282 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990306 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990330 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990407 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.990428 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.991879 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.992209 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:32 crc kubenswrapper[4959]: I1003 13:51:32.993699 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:32.990261 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-l26lm"] Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.003276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.008147 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.009233 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.009908 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.013108 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.013341 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.013505 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.013923 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.014480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.014521 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.014593 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.014652 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.017651 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnpvb\" (UniqueName: \"kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb\") pod \"ceilometer-0\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.027861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.029542 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdfxv\" (UniqueName: \"kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv\") pod \"dnsmasq-dns-55fff446b9-l26lm\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: E1003 13:51:33.033735 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dns-swift-storage-0 kube-api-access-tdfxv ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-55fff446b9-l26lm" podUID="2928d88a-73b9-425d-a180-b7fdbe9ecd88" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.035084 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.036855 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.043552 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4m47\" (UniqueName: \"kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47\") pod \"keystone-bootstrap-gczs6\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.044060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.070294 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.084467 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qnfqp"] Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.086116 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.093171 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.093374 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.093521 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-l2ln8" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.097097 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qnfqp"] Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196802 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196849 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196868 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6bgv\" (UniqueName: \"kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196945 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196965 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.196989 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.197009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.197029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4528s\" (UniqueName: \"kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.197043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.197062 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.197383 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298118 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298575 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4528s\" (UniqueName: \"kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298594 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298618 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298695 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298717 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.298765 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6bgv\" (UniqueName: \"kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.301416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.303217 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.303294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.304160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.304623 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.304723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.307051 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.308161 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.309451 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.318039 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6bgv\" (UniqueName: \"kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv\") pod \"placement-db-sync-qnfqp\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.319477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4528s\" (UniqueName: \"kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s\") pod \"dnsmasq-dns-76fcf4b695-jhcxt\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.469944 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.479092 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnfqp" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.575252 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.597484 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612214 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612336 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdfxv\" (UniqueName: \"kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612513 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612626 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612670 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config\") pod \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\" (UID: \"2928d88a-73b9-425d-a180-b7fdbe9ecd88\") " Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612734 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.612789 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.613119 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.613138 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.613268 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.613672 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config" (OuterVolumeSpecName: "config") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.619777 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv" (OuterVolumeSpecName: "kube-api-access-tdfxv") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "kube-api-access-tdfxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.672529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2928d88a-73b9-425d-a180-b7fdbe9ecd88" (UID: "2928d88a-73b9-425d-a180-b7fdbe9ecd88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.713838 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdfxv\" (UniqueName: \"kubernetes.io/projected/2928d88a-73b9-425d-a180-b7fdbe9ecd88-kube-api-access-tdfxv\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.713871 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.713885 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.713896 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d88a-73b9-425d-a180-b7fdbe9ecd88-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.714576 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593147fe-5977-46ff-8f90-627e3a48142c" path="/var/lib/kubelet/pods/593147fe-5977-46ff-8f90-627e3a48142c/volumes" Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.715616 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gczs6"] Oct 03 13:51:33 crc kubenswrapper[4959]: W1003 13:51:33.734963 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda36f663_16dd_4e00_957b_974aa1e7f47c.slice/crio-e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338 WatchSource:0}: Error finding container e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338: Status 404 returned error can't find the container with id e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338 Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.781704 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:51:33 crc kubenswrapper[4959]: W1003 13:51:33.790202 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f8e7eb1_f105_4b06_aefc_31dc46f989c5.slice/crio-89c838cf8155af8ed73112f3efa0588327a5fdb2c26490cdc5b04edb3d4e8e16 WatchSource:0}: Error finding container 89c838cf8155af8ed73112f3efa0588327a5fdb2c26490cdc5b04edb3d4e8e16: Status 404 returned error can't find the container with id 89c838cf8155af8ed73112f3efa0588327a5fdb2c26490cdc5b04edb3d4e8e16 Oct 03 13:51:33 crc kubenswrapper[4959]: I1003 13:51:33.996556 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:51:34 crc kubenswrapper[4959]: W1003 13:51:34.001765 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd501eb7e_a318_4ecd_8fad_109155ba0763.slice/crio-b3730a68d1e6c5a0d664358618dc86b6184f00b22ac8d1b70d5e5a8acf83482b WatchSource:0}: Error finding container b3730a68d1e6c5a0d664358618dc86b6184f00b22ac8d1b70d5e5a8acf83482b: Status 404 returned error can't find the container with id b3730a68d1e6c5a0d664358618dc86b6184f00b22ac8d1b70d5e5a8acf83482b Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.067305 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qnfqp"] Oct 03 13:51:34 crc kubenswrapper[4959]: W1003 13:51:34.074199 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19f4ff5_e607_4a03_92c0_6365e6d93bf6.slice/crio-8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917 WatchSource:0}: Error finding container 8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917: Status 404 returned error can't find the container with id 8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917 Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.585007 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerStarted","Data":"2005a597211cd6d3b56104e48debb5f573011008cd099ef931a10f3a11e36164"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.585124 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerStarted","Data":"b3730a68d1e6c5a0d664358618dc86b6184f00b22ac8d1b70d5e5a8acf83482b"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.586235 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f8e7eb1-f105-4b06-aefc-31dc46f989c5","Type":"ContainerStarted","Data":"89c838cf8155af8ed73112f3efa0588327a5fdb2c26490cdc5b04edb3d4e8e16"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.587315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnfqp" event={"ID":"c19f4ff5-e607-4a03-92c0-6365e6d93bf6","Type":"ContainerStarted","Data":"8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.589036 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-l26lm" Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.589044 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gczs6" event={"ID":"da36f663-16dd-4e00-957b-974aa1e7f47c","Type":"ContainerStarted","Data":"ce93e24bd637cc785e98cb368132c035f56a137124dfa56c0e7cd03b1a996262"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.589097 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gczs6" event={"ID":"da36f663-16dd-4e00-957b-974aa1e7f47c","Type":"ContainerStarted","Data":"e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338"} Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.618480 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-gczs6" podStartSLOduration=2.618463308 podStartE2EDuration="2.618463308s" podCreationTimestamp="2025-10-03 13:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:34.613718943 +0000 UTC m=+1263.817062350" watchObservedRunningTime="2025-10-03 13:51:34.618463308 +0000 UTC m=+1263.821806725" Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.649076 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-l26lm"] Oct 03 13:51:34 crc kubenswrapper[4959]: I1003 13:51:34.655507 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-l26lm"] Oct 03 13:51:35 crc kubenswrapper[4959]: I1003 13:51:35.106302 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:51:35 crc kubenswrapper[4959]: I1003 13:51:35.602740 4959 generic.go:334] "Generic (PLEG): container finished" podID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerID="2005a597211cd6d3b56104e48debb5f573011008cd099ef931a10f3a11e36164" exitCode=0 Oct 03 13:51:35 crc kubenswrapper[4959]: I1003 13:51:35.603537 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerDied","Data":"2005a597211cd6d3b56104e48debb5f573011008cd099ef931a10f3a11e36164"} Oct 03 13:51:35 crc kubenswrapper[4959]: I1003 13:51:35.698357 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2928d88a-73b9-425d-a180-b7fdbe9ecd88" path="/var/lib/kubelet/pods/2928d88a-73b9-425d-a180-b7fdbe9ecd88/volumes" Oct 03 13:51:36 crc kubenswrapper[4959]: I1003 13:51:36.089183 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:51:36 crc kubenswrapper[4959]: I1003 13:51:36.089343 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:51:36 crc kubenswrapper[4959]: I1003 13:51:36.611346 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerStarted","Data":"c57277a9f005fbbf82e51c493692d3d1b8c379c4ff89dcced7da50fa7ecb7805"} Oct 03 13:51:37 crc kubenswrapper[4959]: I1003 13:51:37.621385 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:37 crc kubenswrapper[4959]: I1003 13:51:37.647290 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" podStartSLOduration=5.647271418 podStartE2EDuration="5.647271418s" podCreationTimestamp="2025-10-03 13:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:37.646487289 +0000 UTC m=+1266.849830706" watchObservedRunningTime="2025-10-03 13:51:37.647271418 +0000 UTC m=+1266.850614855" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.242305 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2a29-account-create-vr7dm"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.243538 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.246052 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.259078 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f4tl\" (UniqueName: \"kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl\") pod \"barbican-2a29-account-create-vr7dm\" (UID: \"c2f39eb9-4894-4b1c-9676-c0e4485932c5\") " pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.260523 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2a29-account-create-vr7dm"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.344071 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cca9-account-create-tp77d"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.345439 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.348428 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.352009 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cca9-account-create-tp77d"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.361550 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f4tl\" (UniqueName: \"kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl\") pod \"barbican-2a29-account-create-vr7dm\" (UID: \"c2f39eb9-4894-4b1c-9676-c0e4485932c5\") " pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.361600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbzfv\" (UniqueName: \"kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv\") pod \"cinder-cca9-account-create-tp77d\" (UID: \"f2448b40-93ae-4b63-b494-fe294dfcdb35\") " pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.385119 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f4tl\" (UniqueName: \"kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl\") pod \"barbican-2a29-account-create-vr7dm\" (UID: \"c2f39eb9-4894-4b1c-9676-c0e4485932c5\") " pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.468326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbzfv\" (UniqueName: \"kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv\") pod \"cinder-cca9-account-create-tp77d\" (UID: \"f2448b40-93ae-4b63-b494-fe294dfcdb35\") " pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.491807 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbzfv\" (UniqueName: \"kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv\") pod \"cinder-cca9-account-create-tp77d\" (UID: \"f2448b40-93ae-4b63-b494-fe294dfcdb35\") " pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.570756 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.647159 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3e00-account-create-ps5fw"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.648546 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.650815 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.655672 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e00-account-create-ps5fw"] Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.668346 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.672148 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44rs2\" (UniqueName: \"kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2\") pod \"neutron-3e00-account-create-ps5fw\" (UID: \"51b37f5d-cfd9-4bad-a768-f9c2d40f279c\") " pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.773958 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44rs2\" (UniqueName: \"kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2\") pod \"neutron-3e00-account-create-ps5fw\" (UID: \"51b37f5d-cfd9-4bad-a768-f9c2d40f279c\") " pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.796457 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44rs2\" (UniqueName: \"kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2\") pod \"neutron-3e00-account-create-ps5fw\" (UID: \"51b37f5d-cfd9-4bad-a768-f9c2d40f279c\") " pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:40 crc kubenswrapper[4959]: I1003 13:51:40.970030 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:43 crc kubenswrapper[4959]: I1003 13:51:43.472557 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:51:43 crc kubenswrapper[4959]: I1003 13:51:43.535254 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:43 crc kubenswrapper[4959]: I1003 13:51:43.535531 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" containerID="cri-o://4791f6cd327460a52e47de9102371966ebd3798e713c07095beacd598e7d3450" gracePeriod=10 Oct 03 13:51:46 crc kubenswrapper[4959]: I1003 13:51:46.213402 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Oct 03 13:51:46 crc kubenswrapper[4959]: I1003 13:51:46.714543 4959 generic.go:334] "Generic (PLEG): container finished" podID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerID="4791f6cd327460a52e47de9102371966ebd3798e713c07095beacd598e7d3450" exitCode=0 Oct 03 13:51:46 crc kubenswrapper[4959]: I1003 13:51:46.714586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" event={"ID":"bcee75a9-6e74-4e2e-b196-623472e8b6ca","Type":"ContainerDied","Data":"4791f6cd327460a52e47de9102371966ebd3798e713c07095beacd598e7d3450"} Oct 03 13:51:50 crc kubenswrapper[4959]: E1003 13:51:50.422990 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Oct 03 13:51:50 crc kubenswrapper[4959]: E1003 13:51:50.423677 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sr56m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-cm7bx_openstack(c41cfce6-87b2-4ff2-8697-e739f3fa1ed7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:51:50 crc kubenswrapper[4959]: E1003 13:51:50.425218 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-cm7bx" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" Oct 03 13:51:50 crc kubenswrapper[4959]: E1003 13:51:50.756114 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-cm7bx" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" Oct 03 13:51:51 crc kubenswrapper[4959]: E1003 13:51:51.678100 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 03 13:51:51 crc kubenswrapper[4959]: E1003 13:51:51.678302 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6bgv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-qnfqp_openstack(c19f4ff5-e607-4a03-92c0-6365e6d93bf6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:51:51 crc kubenswrapper[4959]: E1003 13:51:51.679405 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-qnfqp" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" Oct 03 13:51:51 crc kubenswrapper[4959]: E1003 13:51:51.762983 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-qnfqp" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.354309 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.497822 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.497895 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njprn\" (UniqueName: \"kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.498013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.498046 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.498067 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.498090 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0\") pod \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\" (UID: \"bcee75a9-6e74-4e2e-b196-623472e8b6ca\") " Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.504387 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn" (OuterVolumeSpecName: "kube-api-access-njprn") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "kube-api-access-njprn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.541752 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.542917 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.550554 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config" (OuterVolumeSpecName: "config") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.553641 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.573915 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcee75a9-6e74-4e2e-b196-623472e8b6ca" (UID: "bcee75a9-6e74-4e2e-b196-623472e8b6ca"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600139 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600174 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njprn\" (UniqueName: \"kubernetes.io/projected/bcee75a9-6e74-4e2e-b196-623472e8b6ca-kube-api-access-njprn\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600223 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600234 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600242 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.600250 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcee75a9-6e74-4e2e-b196-623472e8b6ca-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.777996 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" event={"ID":"bcee75a9-6e74-4e2e-b196-623472e8b6ca","Type":"ContainerDied","Data":"c7320710702f8bbb405dfbc439b1a9a26d3af1d04811be66eebcd6d6b80d87c6"} Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.778037 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.778051 4959 scope.go:117] "RemoveContainer" containerID="4791f6cd327460a52e47de9102371966ebd3798e713c07095beacd598e7d3450" Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.780802 4959 generic.go:334] "Generic (PLEG): container finished" podID="da36f663-16dd-4e00-957b-974aa1e7f47c" containerID="ce93e24bd637cc785e98cb368132c035f56a137124dfa56c0e7cd03b1a996262" exitCode=0 Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.780844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gczs6" event={"ID":"da36f663-16dd-4e00-957b-974aa1e7f47c","Type":"ContainerDied","Data":"ce93e24bd637cc785e98cb368132c035f56a137124dfa56c0e7cd03b1a996262"} Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.814465 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.821841 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-jzpz5"] Oct 03 13:51:53 crc kubenswrapper[4959]: I1003 13:51:53.926292 4959 scope.go:117] "RemoveContainer" containerID="1d1fedaf479efcb97775d0f1d8772354aad002fe93020e429df0b895fb0d9726" Oct 03 13:51:54 crc kubenswrapper[4959]: E1003 13:51:54.046813 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 03 13:51:54 crc kubenswrapper[4959]: E1003 13:51:54.047018 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfh667h5cch559hc9h99hd8h667h65h75h584h55h589h679h5f9hd5h674h78h59fh679h54bh674h59dh687h7fh59h544h5h659h5f6hb8hf8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bnpvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(2f8e7eb1-f105-4b06-aefc-31dc46f989c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.336664 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cca9-account-create-tp77d"] Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.376883 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3e00-account-create-ps5fw"] Oct 03 13:51:54 crc kubenswrapper[4959]: W1003 13:51:54.380147 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b37f5d_cfd9_4bad_a768_f9c2d40f279c.slice/crio-08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427 WatchSource:0}: Error finding container 08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427: Status 404 returned error can't find the container with id 08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427 Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.383335 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2a29-account-create-vr7dm"] Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.796266 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cca9-account-create-tp77d" event={"ID":"f2448b40-93ae-4b63-b494-fe294dfcdb35","Type":"ContainerStarted","Data":"a0e8ebea4f33a6079517be7adede19a99d3c556ce12baa9b22ddbebb6d921cd7"} Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.796605 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cca9-account-create-tp77d" event={"ID":"f2448b40-93ae-4b63-b494-fe294dfcdb35","Type":"ContainerStarted","Data":"39ea4d8cae81957e45c2ec3cd87da68369fb72d45545402781cdc4a98eb0e3be"} Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.798754 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e00-account-create-ps5fw" event={"ID":"51b37f5d-cfd9-4bad-a768-f9c2d40f279c","Type":"ContainerStarted","Data":"4d69f154f3479983536c2706de9142a8fb43833ce35669b457243f2cc542426c"} Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.798920 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e00-account-create-ps5fw" event={"ID":"51b37f5d-cfd9-4bad-a768-f9c2d40f279c","Type":"ContainerStarted","Data":"08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427"} Oct 03 13:51:54 crc kubenswrapper[4959]: I1003 13:51:54.801053 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2a29-account-create-vr7dm" event={"ID":"c2f39eb9-4894-4b1c-9676-c0e4485932c5","Type":"ContainerStarted","Data":"79eee14d63092da72f7a3341a9d3b9aec489f138533fc6acf404c06ff10dd6a6"} Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.075521 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.222710 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.222778 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.222886 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.222931 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.223034 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4m47\" (UniqueName: \"kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.223061 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts\") pod \"da36f663-16dd-4e00-957b-974aa1e7f47c\" (UID: \"da36f663-16dd-4e00-957b-974aa1e7f47c\") " Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.228806 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.228820 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts" (OuterVolumeSpecName: "scripts") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.229648 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47" (OuterVolumeSpecName: "kube-api-access-c4m47") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "kube-api-access-c4m47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.229659 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.247547 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data" (OuterVolumeSpecName: "config-data") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.250631 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da36f663-16dd-4e00-957b-974aa1e7f47c" (UID: "da36f663-16dd-4e00-957b-974aa1e7f47c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.325796 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.326276 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.326287 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.326297 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.326306 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4m47\" (UniqueName: \"kubernetes.io/projected/da36f663-16dd-4e00-957b-974aa1e7f47c-kube-api-access-c4m47\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.326315 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da36f663-16dd-4e00-957b-974aa1e7f47c-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.697099 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" path="/var/lib/kubelet/pods/bcee75a9-6e74-4e2e-b196-623472e8b6ca/volumes" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.828807 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gczs6" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.829093 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gczs6" event={"ID":"da36f663-16dd-4e00-957b-974aa1e7f47c","Type":"ContainerDied","Data":"e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338"} Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.829160 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e05e1db59e2ecb304e810cf1f3f9ce8eac167679eaca1aa746addff8cba12338" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.831576 4959 generic.go:334] "Generic (PLEG): container finished" podID="51b37f5d-cfd9-4bad-a768-f9c2d40f279c" containerID="4d69f154f3479983536c2706de9142a8fb43833ce35669b457243f2cc542426c" exitCode=0 Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.832414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e00-account-create-ps5fw" event={"ID":"51b37f5d-cfd9-4bad-a768-f9c2d40f279c","Type":"ContainerDied","Data":"4d69f154f3479983536c2706de9142a8fb43833ce35669b457243f2cc542426c"} Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.838550 4959 generic.go:334] "Generic (PLEG): container finished" podID="c2f39eb9-4894-4b1c-9676-c0e4485932c5" containerID="1ded0051ef4f89932251feaa8d78562d05eed4ad7db125f2a1dfac0af98a16e6" exitCode=0 Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.840158 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2a29-account-create-vr7dm" event={"ID":"c2f39eb9-4894-4b1c-9676-c0e4485932c5","Type":"ContainerDied","Data":"1ded0051ef4f89932251feaa8d78562d05eed4ad7db125f2a1dfac0af98a16e6"} Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.850314 4959 generic.go:334] "Generic (PLEG): container finished" podID="f2448b40-93ae-4b63-b494-fe294dfcdb35" containerID="a0e8ebea4f33a6079517be7adede19a99d3c556ce12baa9b22ddbebb6d921cd7" exitCode=0 Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.850359 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cca9-account-create-tp77d" event={"ID":"f2448b40-93ae-4b63-b494-fe294dfcdb35","Type":"ContainerDied","Data":"a0e8ebea4f33a6079517be7adede19a99d3c556ce12baa9b22ddbebb6d921cd7"} Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.887239 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-gczs6"] Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.892713 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-gczs6"] Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.978158 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-spfmm"] Oct 03 13:51:55 crc kubenswrapper[4959]: E1003 13:51:55.978887 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da36f663-16dd-4e00-957b-974aa1e7f47c" containerName="keystone-bootstrap" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.978912 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="da36f663-16dd-4e00-957b-974aa1e7f47c" containerName="keystone-bootstrap" Oct 03 13:51:55 crc kubenswrapper[4959]: E1003 13:51:55.978931 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.978940 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" Oct 03 13:51:55 crc kubenswrapper[4959]: E1003 13:51:55.978961 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="init" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.978969 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="init" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.979221 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.979247 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="da36f663-16dd-4e00-957b-974aa1e7f47c" containerName="keystone-bootstrap" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.979888 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.982111 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.982335 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.982476 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:51:55 crc kubenswrapper[4959]: I1003 13:51:55.984030 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8gbd" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.000750 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-spfmm"] Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxbl2\" (UniqueName: \"kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039519 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039560 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039588 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.039618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140449 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxbl2\" (UniqueName: \"kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140518 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140538 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140570 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140591 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.140613 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.144788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.145091 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.145312 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.146342 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.147077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.157757 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxbl2\" (UniqueName: \"kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2\") pod \"keystone-bootstrap-spfmm\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.214717 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-jzpz5" podUID="bcee75a9-6e74-4e2e-b196-623472e8b6ca" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.306057 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.784446 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-spfmm"] Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.861753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spfmm" event={"ID":"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2","Type":"ContainerStarted","Data":"ba9165c6148090cd60b9694d96461dec7ada32768fc0405ec9577878692daa5d"} Oct 03 13:51:56 crc kubenswrapper[4959]: I1003 13:51:56.863402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f8e7eb1-f105-4b06-aefc-31dc46f989c5","Type":"ContainerStarted","Data":"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7"} Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.109364 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.260060 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44rs2\" (UniqueName: \"kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2\") pod \"51b37f5d-cfd9-4bad-a768-f9c2d40f279c\" (UID: \"51b37f5d-cfd9-4bad-a768-f9c2d40f279c\") " Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.264563 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2" (OuterVolumeSpecName: "kube-api-access-44rs2") pod "51b37f5d-cfd9-4bad-a768-f9c2d40f279c" (UID: "51b37f5d-cfd9-4bad-a768-f9c2d40f279c"). InnerVolumeSpecName "kube-api-access-44rs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.365787 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44rs2\" (UniqueName: \"kubernetes.io/projected/51b37f5d-cfd9-4bad-a768-f9c2d40f279c-kube-api-access-44rs2\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.422906 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.432741 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.567949 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f4tl\" (UniqueName: \"kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl\") pod \"c2f39eb9-4894-4b1c-9676-c0e4485932c5\" (UID: \"c2f39eb9-4894-4b1c-9676-c0e4485932c5\") " Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.568058 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbzfv\" (UniqueName: \"kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv\") pod \"f2448b40-93ae-4b63-b494-fe294dfcdb35\" (UID: \"f2448b40-93ae-4b63-b494-fe294dfcdb35\") " Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.572687 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl" (OuterVolumeSpecName: "kube-api-access-6f4tl") pod "c2f39eb9-4894-4b1c-9676-c0e4485932c5" (UID: "c2f39eb9-4894-4b1c-9676-c0e4485932c5"). InnerVolumeSpecName "kube-api-access-6f4tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.580864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv" (OuterVolumeSpecName: "kube-api-access-wbzfv") pod "f2448b40-93ae-4b63-b494-fe294dfcdb35" (UID: "f2448b40-93ae-4b63-b494-fe294dfcdb35"). InnerVolumeSpecName "kube-api-access-wbzfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.670388 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f4tl\" (UniqueName: \"kubernetes.io/projected/c2f39eb9-4894-4b1c-9676-c0e4485932c5-kube-api-access-6f4tl\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.670421 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbzfv\" (UniqueName: \"kubernetes.io/projected/f2448b40-93ae-4b63-b494-fe294dfcdb35-kube-api-access-wbzfv\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.694962 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da36f663-16dd-4e00-957b-974aa1e7f47c" path="/var/lib/kubelet/pods/da36f663-16dd-4e00-957b-974aa1e7f47c/volumes" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.874825 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cca9-account-create-tp77d" event={"ID":"f2448b40-93ae-4b63-b494-fe294dfcdb35","Type":"ContainerDied","Data":"39ea4d8cae81957e45c2ec3cd87da68369fb72d45545402781cdc4a98eb0e3be"} Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.874858 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39ea4d8cae81957e45c2ec3cd87da68369fb72d45545402781cdc4a98eb0e3be" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.874862 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cca9-account-create-tp77d" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.882546 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3e00-account-create-ps5fw" event={"ID":"51b37f5d-cfd9-4bad-a768-f9c2d40f279c","Type":"ContainerDied","Data":"08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427"} Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.882579 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08ba8b330131e9b8fd190555222ae27f1b9d0357d0bec7228bbc111626204427" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.882646 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3e00-account-create-ps5fw" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.884454 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2a29-account-create-vr7dm" event={"ID":"c2f39eb9-4894-4b1c-9676-c0e4485932c5","Type":"ContainerDied","Data":"79eee14d63092da72f7a3341a9d3b9aec489f138533fc6acf404c06ff10dd6a6"} Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.884491 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79eee14d63092da72f7a3341a9d3b9aec489f138533fc6acf404c06ff10dd6a6" Oct 03 13:51:57 crc kubenswrapper[4959]: I1003 13:51:57.884504 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2a29-account-create-vr7dm" Oct 03 13:51:58 crc kubenswrapper[4959]: I1003 13:51:58.896145 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spfmm" event={"ID":"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2","Type":"ContainerStarted","Data":"0605bc3c6b468664f6897c5503b23d22fcb8a1132c5793fe99fb04a352189084"} Oct 03 13:51:58 crc kubenswrapper[4959]: I1003 13:51:58.922653 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-spfmm" podStartSLOduration=3.922632746 podStartE2EDuration="3.922632746s" podCreationTimestamp="2025-10-03 13:51:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:51:58.914987701 +0000 UTC m=+1288.118331138" watchObservedRunningTime="2025-10-03 13:51:58.922632746 +0000 UTC m=+1288.125976173" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.569709 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-s9dfg"] Oct 03 13:52:00 crc kubenswrapper[4959]: E1003 13:52:00.570589 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2f39eb9-4894-4b1c-9676-c0e4485932c5" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570606 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2f39eb9-4894-4b1c-9676-c0e4485932c5" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: E1003 13:52:00.570634 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b37f5d-cfd9-4bad-a768-f9c2d40f279c" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570639 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b37f5d-cfd9-4bad-a768-f9c2d40f279c" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: E1003 13:52:00.570662 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2448b40-93ae-4b63-b494-fe294dfcdb35" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570669 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2448b40-93ae-4b63-b494-fe294dfcdb35" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570912 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2448b40-93ae-4b63-b494-fe294dfcdb35" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570930 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2f39eb9-4894-4b1c-9676-c0e4485932c5" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.570944 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b37f5d-cfd9-4bad-a768-f9c2d40f279c" containerName="mariadb-account-create" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.571678 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.575839 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.589775 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zbfhl" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.592377 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s9dfg"] Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.605330 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-d78tr"] Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.606733 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.610790 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.610934 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.611055 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6hgc5" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.632484 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d78tr"] Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722159 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722330 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722820 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mvbz\" (UniqueName: \"kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c266\" (UniqueName: \"kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.722945 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.723043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.723251 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.723298 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c266\" (UniqueName: \"kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824821 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824897 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824918 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824940 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.824985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.825009 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mvbz\" (UniqueName: \"kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.825873 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.845761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.846149 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.846468 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.846749 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.848167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mvbz\" (UniqueName: \"kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz\") pod \"barbican-db-sync-s9dfg\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.848167 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.850499 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.850877 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c266\" (UniqueName: \"kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266\") pod \"cinder-db-sync-d78tr\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.890968 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.918112 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-tdxpk"] Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.919095 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.923565 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jcjmw" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.923803 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.923980 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.924375 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d78tr" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.925618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kldk9\" (UniqueName: \"kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.925678 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.925711 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:00 crc kubenswrapper[4959]: I1003 13:52:00.932316 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tdxpk"] Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.028958 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kldk9\" (UniqueName: \"kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.029033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.029065 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.032972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.039038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.046510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kldk9\" (UniqueName: \"kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9\") pod \"neutron-db-sync-tdxpk\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:01 crc kubenswrapper[4959]: I1003 13:52:01.271687 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:52:06 crc kubenswrapper[4959]: I1003 13:52:06.044962 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:52:06 crc kubenswrapper[4959]: I1003 13:52:06.045616 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:52:14 crc kubenswrapper[4959]: I1003 13:52:14.056262 4959 generic.go:334] "Generic (PLEG): container finished" podID="6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" containerID="0605bc3c6b468664f6897c5503b23d22fcb8a1132c5793fe99fb04a352189084" exitCode=0 Oct 03 13:52:14 crc kubenswrapper[4959]: I1003 13:52:14.056343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spfmm" event={"ID":"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2","Type":"ContainerDied","Data":"0605bc3c6b468664f6897c5503b23d22fcb8a1132c5793fe99fb04a352189084"} Oct 03 13:52:16 crc kubenswrapper[4959]: E1003 13:52:16.696943 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Oct 03 13:52:16 crc kubenswrapper[4959]: E1003 13:52:16.697657 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bnpvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(2f8e7eb1-f105-4b06-aefc-31dc46f989c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.768333 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.915539 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.915870 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.915897 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.915972 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxbl2\" (UniqueName: \"kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.916047 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.916120 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts\") pod \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\" (UID: \"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2\") " Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.923343 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.924666 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts" (OuterVolumeSpecName: "scripts") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.925506 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2" (OuterVolumeSpecName: "kube-api-access-pxbl2") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "kube-api-access-pxbl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.925587 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.946951 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:16 crc kubenswrapper[4959]: I1003 13:52:16.952333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data" (OuterVolumeSpecName: "config-data") pod "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" (UID: "6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018188 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018246 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018257 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018269 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxbl2\" (UniqueName: \"kubernetes.io/projected/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-kube-api-access-pxbl2\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018279 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.018287 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.085978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-spfmm" event={"ID":"6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2","Type":"ContainerDied","Data":"ba9165c6148090cd60b9694d96461dec7ada32768fc0405ec9577878692daa5d"} Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.086025 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba9165c6148090cd60b9694d96461dec7ada32768fc0405ec9577878692daa5d" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.086096 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-spfmm" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.296359 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s9dfg"] Oct 03 13:52:17 crc kubenswrapper[4959]: W1003 13:52:17.302336 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58529f64_d1c9_4feb_ba01_a0cdea44996e.slice/crio-d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1 WatchSource:0}: Error finding container d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1: Status 404 returned error can't find the container with id d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1 Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.304451 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-d78tr"] Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.312963 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tdxpk"] Oct 03 13:52:17 crc kubenswrapper[4959]: W1003 13:52:17.320261 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod440fa439_935e_4cba_8da7_89b2f6a797ea.slice/crio-a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9 WatchSource:0}: Error finding container a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9: Status 404 returned error can't find the container with id a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9 Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.865804 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:52:17 crc kubenswrapper[4959]: E1003 13:52:17.866536 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" containerName="keystone-bootstrap" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.866554 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" containerName="keystone-bootstrap" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.866773 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" containerName="keystone-bootstrap" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.867433 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.869330 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.869583 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s8gbd" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.870551 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.870881 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.870892 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.870923 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.888124 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.935539 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.935584 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.935628 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.935654 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.936043 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.936137 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd8ph\" (UniqueName: \"kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.936317 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:17 crc kubenswrapper[4959]: I1003 13:52:17.936352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037463 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037512 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037588 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037610 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037670 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037687 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd8ph\" (UniqueName: \"kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037719 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.037826 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.042937 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.043119 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.043897 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.044299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.044755 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.048617 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.048779 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.062786 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd8ph\" (UniqueName: \"kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph\") pod \"keystone-5b6c4d45db-cwzjn\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.098086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d78tr" event={"ID":"1433da4c-d3b1-48ea-b62e-e70af51671e5","Type":"ContainerStarted","Data":"1610d5359d738f07bdb8db33ceb61e11ca2a7012f35f93fdf08636df4548a228"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.099500 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cm7bx" event={"ID":"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7","Type":"ContainerStarted","Data":"c0885574e428c640960fc2f2df84effd904ebceef0a6385f1fcac690331dee73"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.103665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnfqp" event={"ID":"c19f4ff5-e607-4a03-92c0-6365e6d93bf6","Type":"ContainerStarted","Data":"bff1a6e521163ac3cf38cf50dfacc21bc264522ed18b6fa7a5c53f6262dbc566"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.105540 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tdxpk" event={"ID":"440fa439-935e-4cba-8da7-89b2f6a797ea","Type":"ContainerStarted","Data":"cfae109f6f3042bf10c7d8abaf5600ce7fac83b01ce2f4be11acd687b892aff1"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.105606 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tdxpk" event={"ID":"440fa439-935e-4cba-8da7-89b2f6a797ea","Type":"ContainerStarted","Data":"a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.107092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s9dfg" event={"ID":"58529f64-d1c9-4feb-ba01-a0cdea44996e","Type":"ContainerStarted","Data":"d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1"} Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.120889 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-cm7bx" podStartSLOduration=2.39339856 podStartE2EDuration="50.120868089s" podCreationTimestamp="2025-10-03 13:51:28 +0000 UTC" firstStartedPulling="2025-10-03 13:51:29.298011007 +0000 UTC m=+1258.501354424" lastFinishedPulling="2025-10-03 13:52:17.025480546 +0000 UTC m=+1306.228823953" observedRunningTime="2025-10-03 13:52:18.118347237 +0000 UTC m=+1307.321690654" watchObservedRunningTime="2025-10-03 13:52:18.120868089 +0000 UTC m=+1307.324211506" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.144696 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-tdxpk" podStartSLOduration=18.144670806 podStartE2EDuration="18.144670806s" podCreationTimestamp="2025-10-03 13:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:18.131178728 +0000 UTC m=+1307.334522145" watchObservedRunningTime="2025-10-03 13:52:18.144670806 +0000 UTC m=+1307.348014223" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.154977 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qnfqp" podStartSLOduration=2.195367479 podStartE2EDuration="45.154958426s" podCreationTimestamp="2025-10-03 13:51:33 +0000 UTC" firstStartedPulling="2025-10-03 13:51:34.076715122 +0000 UTC m=+1263.280058539" lastFinishedPulling="2025-10-03 13:52:17.036306069 +0000 UTC m=+1306.239649486" observedRunningTime="2025-10-03 13:52:18.148271903 +0000 UTC m=+1307.351615330" watchObservedRunningTime="2025-10-03 13:52:18.154958426 +0000 UTC m=+1307.358301843" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.187773 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:18 crc kubenswrapper[4959]: I1003 13:52:18.671782 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:52:18 crc kubenswrapper[4959]: W1003 13:52:18.686168 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf51595d7_80bf_4436_9429_9f0da918850d.slice/crio-e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd WatchSource:0}: Error finding container e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd: Status 404 returned error can't find the container with id e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd Oct 03 13:52:19 crc kubenswrapper[4959]: I1003 13:52:19.117813 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b6c4d45db-cwzjn" event={"ID":"f51595d7-80bf-4436-9429-9f0da918850d","Type":"ContainerStarted","Data":"e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd"} Oct 03 13:52:20 crc kubenswrapper[4959]: I1003 13:52:20.131893 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b6c4d45db-cwzjn" event={"ID":"f51595d7-80bf-4436-9429-9f0da918850d","Type":"ContainerStarted","Data":"964de272e99fd5d3325f928f234440a05a23315ec4d4c92db1d5a71e50facdcf"} Oct 03 13:52:20 crc kubenswrapper[4959]: I1003 13:52:20.132275 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:20 crc kubenswrapper[4959]: I1003 13:52:20.157503 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b6c4d45db-cwzjn" podStartSLOduration=3.15748135 podStartE2EDuration="3.15748135s" podCreationTimestamp="2025-10-03 13:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:20.154185131 +0000 UTC m=+1309.357528578" watchObservedRunningTime="2025-10-03 13:52:20.15748135 +0000 UTC m=+1309.360824767" Oct 03 13:52:21 crc kubenswrapper[4959]: I1003 13:52:21.140728 4959 generic.go:334] "Generic (PLEG): container finished" podID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" containerID="bff1a6e521163ac3cf38cf50dfacc21bc264522ed18b6fa7a5c53f6262dbc566" exitCode=0 Oct 03 13:52:21 crc kubenswrapper[4959]: I1003 13:52:21.141591 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnfqp" event={"ID":"c19f4ff5-e607-4a03-92c0-6365e6d93bf6","Type":"ContainerDied","Data":"bff1a6e521163ac3cf38cf50dfacc21bc264522ed18b6fa7a5c53f6262dbc566"} Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.349032 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnfqp" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482454 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs\") pod \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482517 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts\") pod \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482578 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle\") pod \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482651 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6bgv\" (UniqueName: \"kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv\") pod \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482697 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data\") pod \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\" (UID: \"c19f4ff5-e607-4a03-92c0-6365e6d93bf6\") " Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.482986 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs" (OuterVolumeSpecName: "logs") pod "c19f4ff5-e607-4a03-92c0-6365e6d93bf6" (UID: "c19f4ff5-e607-4a03-92c0-6365e6d93bf6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.483590 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.490564 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts" (OuterVolumeSpecName: "scripts") pod "c19f4ff5-e607-4a03-92c0-6365e6d93bf6" (UID: "c19f4ff5-e607-4a03-92c0-6365e6d93bf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.501539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv" (OuterVolumeSpecName: "kube-api-access-g6bgv") pod "c19f4ff5-e607-4a03-92c0-6365e6d93bf6" (UID: "c19f4ff5-e607-4a03-92c0-6365e6d93bf6"). InnerVolumeSpecName "kube-api-access-g6bgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.526461 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c19f4ff5-e607-4a03-92c0-6365e6d93bf6" (UID: "c19f4ff5-e607-4a03-92c0-6365e6d93bf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.531073 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data" (OuterVolumeSpecName: "config-data") pod "c19f4ff5-e607-4a03-92c0-6365e6d93bf6" (UID: "c19f4ff5-e607-4a03-92c0-6365e6d93bf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.585086 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6bgv\" (UniqueName: \"kubernetes.io/projected/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-kube-api-access-g6bgv\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.585120 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.585129 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:25 crc kubenswrapper[4959]: I1003 13:52:25.585138 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19f4ff5-e607-4a03-92c0-6365e6d93bf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.186092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qnfqp" event={"ID":"c19f4ff5-e607-4a03-92c0-6365e6d93bf6","Type":"ContainerDied","Data":"8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917"} Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.186162 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b53271e9caea63da102da8b525e13c0424dc540d004fedb990930932a5cd917" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.186176 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qnfqp" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.457407 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:52:26 crc kubenswrapper[4959]: E1003 13:52:26.457748 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" containerName="placement-db-sync" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.457760 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" containerName="placement-db-sync" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.457945 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" containerName="placement-db-sync" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.458833 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.461535 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.461620 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.468247 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.470607 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.470636 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-l2ln8" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.476347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.601660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.601790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.601984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.602131 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.602217 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbt7t\" (UniqueName: \"kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.602267 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.602314 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.704644 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.704726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.704822 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.704935 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.704972 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbt7t\" (UniqueName: \"kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.705034 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.705085 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.705097 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.708999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.709509 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.709587 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.710331 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.710761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.724921 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbt7t\" (UniqueName: \"kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t\") pod \"placement-86678496f6-dwd5r\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:26 crc kubenswrapper[4959]: I1003 13:52:26.776826 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:33 crc kubenswrapper[4959]: E1003 13:52:33.631776 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 03 13:52:33 crc kubenswrapper[4959]: E1003 13:52:33.632348 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2c266,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-d78tr_openstack(1433da4c-d3b1-48ea-b62e-e70af51671e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:52:33 crc kubenswrapper[4959]: E1003 13:52:33.633661 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-d78tr" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" Oct 03 13:52:34 crc kubenswrapper[4959]: E1003 13:52:34.263062 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-d78tr" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.044580 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.044964 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.045020 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.045825 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.045892 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77" gracePeriod=600 Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.284646 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77" exitCode=0 Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.284713 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77"} Oct 03 13:52:36 crc kubenswrapper[4959]: I1003 13:52:36.284774 4959 scope.go:117] "RemoveContainer" containerID="c09eebf4d04f46983e56a152d8e3536a6344578e226f4757ffb462a20e528694" Oct 03 13:52:39 crc kubenswrapper[4959]: E1003 13:52:39.331004 4959 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 03 13:52:39 crc kubenswrapper[4959]: E1003 13:52:39.331602 4959 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bnpvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(2f8e7eb1-f105-4b06-aefc-31dc46f989c5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:52:39 crc kubenswrapper[4959]: E1003 13:52:39.333174 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" Oct 03 13:52:39 crc kubenswrapper[4959]: I1003 13:52:39.750402 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.328404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s9dfg" event={"ID":"58529f64-d1c9-4feb-ba01-a0cdea44996e","Type":"ContainerStarted","Data":"a88bf5ca4e0c7d6091ecce6d3abd4e46c1d584111080db1a1653bc5d9ffebe87"} Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.331041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c"} Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.333604 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" containerName="ceilometer-notification-agent" containerID="cri-o://ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7" gracePeriod=30 Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.334286 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerStarted","Data":"617f60df069f8b127236c30ef29dbcf5ad86abf5cef1969b4f6275e6d0b31ca5"} Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.334365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerStarted","Data":"e2a2488abba43c5130cc5e66793ff02a224a83cad1a37062cb627eb282992260"} Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.334390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerStarted","Data":"0640b5671c1ba9b42e33af591d1536c6dae27bd4d2642a6ca2f4607601a4d974"} Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.334549 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.334581 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.391785 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-s9dfg" podStartSLOduration=18.37561526 podStartE2EDuration="40.391765395s" podCreationTimestamp="2025-10-03 13:52:00 +0000 UTC" firstStartedPulling="2025-10-03 13:52:17.304931688 +0000 UTC m=+1306.508275105" lastFinishedPulling="2025-10-03 13:52:39.321081783 +0000 UTC m=+1328.524425240" observedRunningTime="2025-10-03 13:52:40.353459086 +0000 UTC m=+1329.556802503" watchObservedRunningTime="2025-10-03 13:52:40.391765395 +0000 UTC m=+1329.595108832" Oct 03 13:52:40 crc kubenswrapper[4959]: I1003 13:52:40.430573 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-86678496f6-dwd5r" podStartSLOduration=14.430553177 podStartE2EDuration="14.430553177s" podCreationTimestamp="2025-10-03 13:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:40.42534641 +0000 UTC m=+1329.628689837" watchObservedRunningTime="2025-10-03 13:52:40.430553177 +0000 UTC m=+1329.633896594" Oct 03 13:52:43 crc kubenswrapper[4959]: I1003 13:52:43.361352 4959 generic.go:334] "Generic (PLEG): container finished" podID="58529f64-d1c9-4feb-ba01-a0cdea44996e" containerID="a88bf5ca4e0c7d6091ecce6d3abd4e46c1d584111080db1a1653bc5d9ffebe87" exitCode=0 Oct 03 13:52:43 crc kubenswrapper[4959]: I1003 13:52:43.361695 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s9dfg" event={"ID":"58529f64-d1c9-4feb-ba01-a0cdea44996e","Type":"ContainerDied","Data":"a88bf5ca4e0c7d6091ecce6d3abd4e46c1d584111080db1a1653bc5d9ffebe87"} Oct 03 13:52:43 crc kubenswrapper[4959]: I1003 13:52:43.364302 4959 generic.go:334] "Generic (PLEG): container finished" podID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" containerID="c0885574e428c640960fc2f2df84effd904ebceef0a6385f1fcac690331dee73" exitCode=0 Oct 03 13:52:43 crc kubenswrapper[4959]: I1003 13:52:43.364335 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cm7bx" event={"ID":"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7","Type":"ContainerDied","Data":"c0885574e428c640960fc2f2df84effd904ebceef0a6385f1fcac690331dee73"} Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.292334 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.377426 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" containerID="ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7" exitCode=0 Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.377508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f8e7eb1-f105-4b06-aefc-31dc46f989c5","Type":"ContainerDied","Data":"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7"} Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.377593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f8e7eb1-f105-4b06-aefc-31dc46f989c5","Type":"ContainerDied","Data":"89c838cf8155af8ed73112f3efa0588327a5fdb2c26490cdc5b04edb3d4e8e16"} Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.377631 4959 scope.go:117] "RemoveContainer" containerID="ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.377767 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408257 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408337 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408387 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408472 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408654 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnpvb\" (UniqueName: \"kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408695 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408733 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml\") pod \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\" (UID: \"2f8e7eb1-f105-4b06-aefc-31dc46f989c5\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.408888 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.409174 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.409533 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.409553 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.414333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.414368 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts" (OuterVolumeSpecName: "scripts") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.415019 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb" (OuterVolumeSpecName: "kube-api-access-bnpvb") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "kube-api-access-bnpvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.416376 4959 scope.go:117] "RemoveContainer" containerID="ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7" Oct 03 13:52:44 crc kubenswrapper[4959]: E1003 13:52:44.416984 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7\": container with ID starting with ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7 not found: ID does not exist" containerID="ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.417035 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7"} err="failed to get container status \"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7\": rpc error: code = NotFound desc = could not find container \"ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7\": container with ID starting with ff8022110b9de28ad4abf36b4d5b744d959dae8032883a146dfa303869460fd7 not found: ID does not exist" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.438145 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data" (OuterVolumeSpecName: "config-data") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.439392 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f8e7eb1-f105-4b06-aefc-31dc46f989c5" (UID: "2f8e7eb1-f105-4b06-aefc-31dc46f989c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.511007 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.511041 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.511050 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.511058 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnpvb\" (UniqueName: \"kubernetes.io/projected/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-kube-api-access-bnpvb\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.511068 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f8e7eb1-f105-4b06-aefc-31dc46f989c5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.610302 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.713728 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mvbz\" (UniqueName: \"kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz\") pod \"58529f64-d1c9-4feb-ba01-a0cdea44996e\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.713765 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle\") pod \"58529f64-d1c9-4feb-ba01-a0cdea44996e\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.713837 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data\") pod \"58529f64-d1c9-4feb-ba01-a0cdea44996e\" (UID: \"58529f64-d1c9-4feb-ba01-a0cdea44996e\") " Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.722176 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz" (OuterVolumeSpecName: "kube-api-access-6mvbz") pod "58529f64-d1c9-4feb-ba01-a0cdea44996e" (UID: "58529f64-d1c9-4feb-ba01-a0cdea44996e"). InnerVolumeSpecName "kube-api-access-6mvbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.727823 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "58529f64-d1c9-4feb-ba01-a0cdea44996e" (UID: "58529f64-d1c9-4feb-ba01-a0cdea44996e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.745832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58529f64-d1c9-4feb-ba01-a0cdea44996e" (UID: "58529f64-d1c9-4feb-ba01-a0cdea44996e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.753384 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.761432 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.771305 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:52:44 crc kubenswrapper[4959]: E1003 13:52:44.771763 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" containerName="ceilometer-notification-agent" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.771788 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" containerName="ceilometer-notification-agent" Oct 03 13:52:44 crc kubenswrapper[4959]: E1003 13:52:44.771821 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58529f64-d1c9-4feb-ba01-a0cdea44996e" containerName="barbican-db-sync" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.771831 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="58529f64-d1c9-4feb-ba01-a0cdea44996e" containerName="barbican-db-sync" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.772037 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="58529f64-d1c9-4feb-ba01-a0cdea44996e" containerName="barbican-db-sync" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.772069 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" containerName="ceilometer-notification-agent" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.773978 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.777264 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.777564 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.780435 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.815619 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mvbz\" (UniqueName: \"kubernetes.io/projected/58529f64-d1c9-4feb-ba01-a0cdea44996e-kube-api-access-6mvbz\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.815646 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.815655 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/58529f64-d1c9-4feb-ba01-a0cdea44996e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916559 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s95b\" (UniqueName: \"kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916639 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916697 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916713 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:44 crc kubenswrapper[4959]: I1003 13:52:44.916928 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.018950 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019001 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s95b\" (UniqueName: \"kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019058 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019078 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019122 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019152 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.019871 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.020490 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.024427 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.024572 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.024715 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.025235 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.039783 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s95b\" (UniqueName: \"kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b\") pod \"ceilometer-0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.097492 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.107600 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cm7bx" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.221776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr56m\" (UniqueName: \"kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m\") pod \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.221916 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data\") pod \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.221967 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle\") pod \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.222033 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data\") pod \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\" (UID: \"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7\") " Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.225521 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" (UID: "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.227072 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m" (OuterVolumeSpecName: "kube-api-access-sr56m") pod "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" (UID: "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7"). InnerVolumeSpecName "kube-api-access-sr56m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.277477 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" (UID: "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.284794 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data" (OuterVolumeSpecName: "config-data") pod "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" (UID: "c41cfce6-87b2-4ff2-8697-e739f3fa1ed7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.324810 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.324846 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.324856 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.324864 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr56m\" (UniqueName: \"kubernetes.io/projected/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7-kube-api-access-sr56m\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.389873 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s9dfg" event={"ID":"58529f64-d1c9-4feb-ba01-a0cdea44996e","Type":"ContainerDied","Data":"d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1"} Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.390116 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d31889d2b12617b89e413b2ff514f148f2e7a40e66b1dba7e5aff0b9f21d80e1" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.390533 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s9dfg" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.398866 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cm7bx" event={"ID":"c41cfce6-87b2-4ff2-8697-e739f3fa1ed7","Type":"ContainerDied","Data":"e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811"} Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.398903 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9205bdebe284205a8eea67b1e2352f01ade79b3efa349e75a2b6faae394e811" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.399076 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cm7bx" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.584803 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:52:45 crc kubenswrapper[4959]: W1003 13:52:45.585631 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779685e0_3882_4dbe_ad13_df648b89d1a0.slice/crio-78a375752b390c039898849196854c58e23d561421eca4e1b6131eeb828ce19d WatchSource:0}: Error finding container 78a375752b390c039898849196854c58e23d561421eca4e1b6131eeb828ce19d: Status 404 returned error can't find the container with id 78a375752b390c039898849196854c58e23d561421eca4e1b6131eeb828ce19d Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.591338 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:52:45 crc kubenswrapper[4959]: E1003 13:52:45.591681 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" containerName="glance-db-sync" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.591695 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" containerName="glance-db-sync" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.591846 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" containerName="glance-db-sync" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.592675 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.594689 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zbfhl" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.595074 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.599834 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.612434 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.628177 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.641751 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.646601 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.680329 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.718408 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f8e7eb1-f105-4b06-aefc-31dc46f989c5" path="/var/lib/kubelet/pods/2f8e7eb1-f105-4b06-aefc-31dc46f989c5/volumes" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.718960 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.720726 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.720832 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.731355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.731426 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8wr\" (UniqueName: \"kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.731451 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.731474 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.731525 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836574 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836646 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmjk\" (UniqueName: \"kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836685 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836715 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836750 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836813 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdj2m\" (UniqueName: \"kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836846 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836874 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8wr\" (UniqueName: \"kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836897 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836929 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836951 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.836978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.837008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.837033 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.837059 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.837085 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.837599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.849275 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.851884 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.853727 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.870201 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8wr\" (UniqueName: \"kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr\") pod \"barbican-keystone-listener-7f558c656-kjrrw\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.923478 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939046 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmjk\" (UniqueName: \"kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939095 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939119 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939186 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdj2m\" (UniqueName: \"kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939695 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939751 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939782 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939802 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939829 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939849 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.939921 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.940152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.940504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.940658 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.940870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.941008 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.948154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.950024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.968851 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.971895 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.973668 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.988483 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 03 13:52:45 crc kubenswrapper[4959]: I1003 13:52:45.994946 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdj2m\" (UniqueName: \"kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m\") pod \"dnsmasq-dns-7979dc8455-vj2l5\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.000924 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.008055 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmjk\" (UniqueName: \"kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk\") pod \"barbican-worker-54b8ffdcf-7hs45\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.044573 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.144049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.144095 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.144112 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.144137 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.144205 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sdzk\" (UniqueName: \"kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.158638 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.234048 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.236852 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248003 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248059 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248081 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248112 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248168 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sdzk\" (UniqueName: \"kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.248517 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.255990 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.256134 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.257395 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.268515 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.271852 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.318062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sdzk\" (UniqueName: \"kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk\") pod \"barbican-api-684fb7ddcd-zvzm9\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353264 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353374 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353407 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353445 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353490 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92rzf\" (UniqueName: \"kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.353526 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.419131 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerStarted","Data":"78a375752b390c039898849196854c58e23d561421eca4e1b6131eeb828ce19d"} Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455517 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455584 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92rzf\" (UniqueName: \"kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455637 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.455710 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.456124 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.456238 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.456801 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.457505 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.458157 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.473053 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92rzf\" (UniqueName: \"kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf\") pod \"dnsmasq-dns-59d5ff467f-wj4hv\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.492727 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.601759 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.602505 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:46 crc kubenswrapper[4959]: W1003 13:52:46.611261 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ec784f6_4b46_435e_af12_e899033f8348.slice/crio-16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c WatchSource:0}: Error finding container 16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c: Status 404 returned error can't find the container with id 16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.707776 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:46 crc kubenswrapper[4959]: W1003 13:52:46.887715 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaa51dc6_b33c_41fc_9829_e4d20da77df3.slice/crio-90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04 WatchSource:0}: Error finding container 90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04: Status 404 returned error can't find the container with id 90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04 Oct 03 13:52:46 crc kubenswrapper[4959]: I1003 13:52:46.900169 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.004075 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:52:47 crc kubenswrapper[4959]: W1003 13:52:47.032622 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb074b3a0_c821_49c6_bbef_a95e02e45fb2.slice/crio-f98018770d73ca1a06941c0b77f8ce0fd0bdcaf2f9421172d872f62b005e060d WatchSource:0}: Error finding container f98018770d73ca1a06941c0b77f8ce0fd0bdcaf2f9421172d872f62b005e060d: Status 404 returned error can't find the container with id f98018770d73ca1a06941c0b77f8ce0fd0bdcaf2f9421172d872f62b005e060d Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.114316 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.169370 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.172913 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.177299 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-blm84" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.177362 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.177305 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.196398 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.296243 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.298061 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.301171 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.306259 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.381318 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.381381 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.381425 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.381469 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.382040 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.382172 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.382238 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w87dm\" (UniqueName: \"kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.430135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerStarted","Data":"f98018770d73ca1a06941c0b77f8ce0fd0bdcaf2f9421172d872f62b005e060d"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.431975 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerStarted","Data":"90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.433257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerStarted","Data":"16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.447637 4959 generic.go:334] "Generic (PLEG): container finished" podID="7e4c8416-acd9-4e2b-aa50-f42787b6b617" containerID="c51a8e50a698472b09eb87b17561f2b4e00981b1a8a5f13ef810c49d99b9abf6" exitCode=0 Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.447733 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" event={"ID":"7e4c8416-acd9-4e2b-aa50-f42787b6b617","Type":"ContainerDied","Data":"c51a8e50a698472b09eb87b17561f2b4e00981b1a8a5f13ef810c49d99b9abf6"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.447765 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" event={"ID":"7e4c8416-acd9-4e2b-aa50-f42787b6b617","Type":"ContainerStarted","Data":"5e96416244c3d3fb5dfbacf2fa3ca1f10faf31022f6a0f3609fde5f02633c9dd"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.451402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" event={"ID":"e5e4fb74-f494-4b58-8ea2-1af811580474","Type":"ContainerStarted","Data":"c1d4f567be5ac79300618904629545474c97feab4e5c4ad38b36a213a0334a3f"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.455830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerStarted","Data":"d1a6cf72b4294e6230e6e22e8ebf1e91d9bdb34958530232770c3565f560f85b"} Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483569 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483648 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483696 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg7cb\" (UniqueName: \"kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483760 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483790 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483813 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483828 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483848 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483893 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483911 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w87dm\" (UniqueName: \"kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483931 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.483978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.484382 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.484813 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.484867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.494858 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.503894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.504616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.511176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w87dm\" (UniqueName: \"kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.575302 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585070 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585212 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585279 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg7cb\" (UniqueName: \"kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585316 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585366 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585395 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.585424 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.586301 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.603561 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.611860 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.612176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.612384 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.612476 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.616790 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg7cb\" (UniqueName: \"kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.631595 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.713161 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.723006 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.725653 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890065 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890134 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890232 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdj2m\" (UniqueName: \"kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890332 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890376 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.890422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb\") pod \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\" (UID: \"7e4c8416-acd9-4e2b-aa50-f42787b6b617\") " Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.895688 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m" (OuterVolumeSpecName: "kube-api-access-qdj2m") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "kube-api-access-qdj2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.917074 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config" (OuterVolumeSpecName: "config") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.917080 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.918630 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.923042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.924904 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e4c8416-acd9-4e2b-aa50-f42787b6b617" (UID: "7e4c8416-acd9-4e2b-aa50-f42787b6b617"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.992981 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.993032 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.993044 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.993053 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.993061 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8416-acd9-4e2b-aa50-f42787b6b617-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:47 crc kubenswrapper[4959]: I1003 13:52:47.993069 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdj2m\" (UniqueName: \"kubernetes.io/projected/7e4c8416-acd9-4e2b-aa50-f42787b6b617-kube-api-access-qdj2m\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.466798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerStarted","Data":"004a0ee4f59f60160da67b71b5e1962f4e4473c3ff1b2699c4ef12fcaa6e500d"} Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.467138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerStarted","Data":"be3ac4d3f6020be7c862df77e04e33b60d571c907f1325c96ff42d1be4628d4f"} Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.467227 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.467258 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.468976 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" event={"ID":"7e4c8416-acd9-4e2b-aa50-f42787b6b617","Type":"ContainerDied","Data":"5e96416244c3d3fb5dfbacf2fa3ca1f10faf31022f6a0f3609fde5f02633c9dd"} Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.469016 4959 scope.go:117] "RemoveContainer" containerID="c51a8e50a698472b09eb87b17561f2b4e00981b1a8a5f13ef810c49d99b9abf6" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.469117 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7979dc8455-vj2l5" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.476060 4959 generic.go:334] "Generic (PLEG): container finished" podID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerID="e8e43310c8bff99fcaeaf193ddceb252a39706c5515adc44ff056f2f6182f712" exitCode=0 Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.476114 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" event={"ID":"e5e4fb74-f494-4b58-8ea2-1af811580474","Type":"ContainerDied","Data":"e8e43310c8bff99fcaeaf193ddceb252a39706c5515adc44ff056f2f6182f712"} Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.478208 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerStarted","Data":"116ab8d592eea1dea5fa83f80f343889469bd35b63bf2d67c31a6881bdda29c1"} Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.498621 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podStartSLOduration=3.498602713 podStartE2EDuration="3.498602713s" podCreationTimestamp="2025-10-03 13:52:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:48.492507006 +0000 UTC m=+1337.695850443" watchObservedRunningTime="2025-10-03 13:52:48.498602713 +0000 UTC m=+1337.701946130" Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.558651 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:48 crc kubenswrapper[4959]: I1003 13:52:48.564265 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7979dc8455-vj2l5"] Oct 03 13:52:49 crc kubenswrapper[4959]: I1003 13:52:49.702154 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e4c8416-acd9-4e2b-aa50-f42787b6b617" path="/var/lib/kubelet/pods/7e4c8416-acd9-4e2b-aa50-f42787b6b617/volumes" Oct 03 13:52:49 crc kubenswrapper[4959]: I1003 13:52:49.850587 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:52:50 crc kubenswrapper[4959]: I1003 13:52:50.693817 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:52:50 crc kubenswrapper[4959]: I1003 13:52:50.766061 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:52:51 crc kubenswrapper[4959]: I1003 13:52:51.651874 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.347279 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:52:52 crc kubenswrapper[4959]: E1003 13:52:52.347637 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e4c8416-acd9-4e2b-aa50-f42787b6b617" containerName="init" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.347648 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e4c8416-acd9-4e2b-aa50-f42787b6b617" containerName="init" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.347807 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e4c8416-acd9-4e2b-aa50-f42787b6b617" containerName="init" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.348691 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.351267 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.351288 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.360683 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.480867 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481159 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481374 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481506 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ck75\" (UniqueName: \"kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481597 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.481700 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.583422 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.583687 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.583761 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.583862 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.583966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.584090 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ck75\" (UniqueName: \"kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.584212 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.584721 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.589746 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.589884 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.591489 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.592260 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.593465 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.607827 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ck75\" (UniqueName: \"kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75\") pod \"barbican-api-677d776894-ljrrj\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:52 crc kubenswrapper[4959]: I1003 13:52:52.681650 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.373876 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.375465 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.378661 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.379371 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xrwhf" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.379422 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.398964 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.526209 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p26j\" (UniqueName: \"kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.526361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.526398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.526475 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.618787 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: E1003 13:52:54.619517 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-5p26j openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="de67c342-4684-4cce-b044-92ba70e1ce26" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.626387 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.627571 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.627609 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.627661 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.627698 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p26j\" (UniqueName: \"kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.628601 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: E1003 13:52:54.629214 4959 projected.go:194] Error preparing data for projected volume kube-api-access-5p26j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 03 13:52:54 crc kubenswrapper[4959]: E1003 13:52:54.629347 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j podName:de67c342-4684-4cce-b044-92ba70e1ce26 nodeName:}" failed. No retries permitted until 2025-10-03 13:52:55.129329677 +0000 UTC m=+1344.332673094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5p26j" (UniqueName: "kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j") pod "openstackclient" (UID: "de67c342-4684-4cce-b044-92ba70e1ce26") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.637485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.649298 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.650231 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.667774 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.668266 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.830920 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.831016 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lw6\" (UniqueName: \"kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.831082 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.831143 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.933072 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.934096 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.934019 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.934783 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.935006 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5lw6\" (UniqueName: \"kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.937731 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.953302 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.954543 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:54 crc kubenswrapper[4959]: I1003 13:52:54.955425 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5lw6\" (UniqueName: \"kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6\") pod \"openstackclient\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " pod="openstack/openstackclient" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.048075 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.138958 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p26j\" (UniqueName: \"kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j\") pod \"openstackclient\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " pod="openstack/openstackclient" Oct 03 13:52:55 crc kubenswrapper[4959]: E1003 13:52:55.140936 4959 projected.go:194] Error preparing data for projected volume kube-api-access-5p26j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (de67c342-4684-4cce-b044-92ba70e1ce26) does not match the UID in record. The object might have been deleted and then recreated Oct 03 13:52:55 crc kubenswrapper[4959]: E1003 13:52:55.141009 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j podName:de67c342-4684-4cce-b044-92ba70e1ce26 nodeName:}" failed. No retries permitted until 2025-10-03 13:52:56.140991584 +0000 UTC m=+1345.344335001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-5p26j" (UniqueName: "kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j") pod "openstackclient" (UID: "de67c342-4684-4cce-b044-92ba70e1ce26") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (de67c342-4684-4cce-b044-92ba70e1ce26) does not match the UID in record. The object might have been deleted and then recreated Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.584761 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.585671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerStarted","Data":"e6c648a2da50214bcfce64fe3f34cc572ea4c8df28649f9ef5c9193e8e513896"} Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.594984 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="de67c342-4684-4cce-b044-92ba70e1ce26" podUID="a7088287-7362-42f5-987c-be9f700c50cb" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.610506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.650654 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.753263 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret\") pod \"de67c342-4684-4cce-b044-92ba70e1ce26\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.753693 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle\") pod \"de67c342-4684-4cce-b044-92ba70e1ce26\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.753719 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config\") pod \"de67c342-4684-4cce-b044-92ba70e1ce26\" (UID: \"de67c342-4684-4cce-b044-92ba70e1ce26\") " Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.754168 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p26j\" (UniqueName: \"kubernetes.io/projected/de67c342-4684-4cce-b044-92ba70e1ce26-kube-api-access-5p26j\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.754739 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "de67c342-4684-4cce-b044-92ba70e1ce26" (UID: "de67c342-4684-4cce-b044-92ba70e1ce26"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.758493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "de67c342-4684-4cce-b044-92ba70e1ce26" (UID: "de67c342-4684-4cce-b044-92ba70e1ce26"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.760835 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de67c342-4684-4cce-b044-92ba70e1ce26" (UID: "de67c342-4684-4cce-b044-92ba70e1ce26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:52:55 crc kubenswrapper[4959]: W1003 13:52:55.774448 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod429beb30_1784_4027_a0f0_230b9e864dcf.slice/crio-21f62d26ac6df23d7b61c9030bf20759f98cdcfcf15a1c24b12f35cfbb26f4ee WatchSource:0}: Error finding container 21f62d26ac6df23d7b61c9030bf20759f98cdcfcf15a1c24b12f35cfbb26f4ee: Status 404 returned error can't find the container with id 21f62d26ac6df23d7b61c9030bf20759f98cdcfcf15a1c24b12f35cfbb26f4ee Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.856215 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.856245 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:55 crc kubenswrapper[4959]: I1003 13:52:55.856259 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de67c342-4684-4cce-b044-92ba70e1ce26-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.055363 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.370307 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:52:56 crc kubenswrapper[4959]: W1003 13:52:56.384706 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7088287_7362_42f5_987c_be9f700c50cb.slice/crio-036b69fd5344f554d8546fc1a390c12d52dcf6cf1d7d2b217491dd296b9ca046 WatchSource:0}: Error finding container 036b69fd5344f554d8546fc1a390c12d52dcf6cf1d7d2b217491dd296b9ca046: Status 404 returned error can't find the container with id 036b69fd5344f554d8546fc1a390c12d52dcf6cf1d7d2b217491dd296b9ca046 Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.619466 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a7088287-7362-42f5-987c-be9f700c50cb","Type":"ContainerStarted","Data":"036b69fd5344f554d8546fc1a390c12d52dcf6cf1d7d2b217491dd296b9ca046"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.633278 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" event={"ID":"e5e4fb74-f494-4b58-8ea2-1af811580474","Type":"ContainerStarted","Data":"9e99639ccaed7b57e2b9ffe2295046f5e3e386fac427e02a3d1ce7f2f3732062"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.634596 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.639706 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerStarted","Data":"282699a695dc3c74712766e4667ffd36463c08f2553c45fa465d271ca4ae5b72"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.648934 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerStarted","Data":"1f8f42d263cd44bb7f3d741af35c4900133542e621536ce527da52ab82dbee5c"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.648989 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerStarted","Data":"b1768aacb21314d2a0749c15d34ef9d67cc08b934e893da49b8d325044b8c354"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.651685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerStarted","Data":"21f62d26ac6df23d7b61c9030bf20759f98cdcfcf15a1c24b12f35cfbb26f4ee"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.675655 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" podStartSLOduration=10.675635155 podStartE2EDuration="10.675635155s" podCreationTimestamp="2025-10-03 13:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:56.662498606 +0000 UTC m=+1345.865842033" watchObservedRunningTime="2025-10-03 13:52:56.675635155 +0000 UTC m=+1345.878978572" Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.700692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerStarted","Data":"d5033b93560574e923577b90dc0e6761b4304b82f8fedaf8d5ae2041e3f74a66"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.726519 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.729609 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerStarted","Data":"023f6f41efd8b70bb5c4bc20ca5e840eaa4f8d065b77a0f728e8d0ecf836a688"} Oct 03 13:52:56 crc kubenswrapper[4959]: I1003 13:52:56.750801 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="de67c342-4684-4cce-b044-92ba70e1ce26" podUID="a7088287-7362-42f5-987c-be9f700c50cb" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.755094 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de67c342-4684-4cce-b044-92ba70e1ce26" path="/var/lib/kubelet/pods/de67c342-4684-4cce-b044-92ba70e1ce26/volumes" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.807144 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerStarted","Data":"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.836136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d78tr" event={"ID":"1433da4c-d3b1-48ea-b62e-e70af51671e5","Type":"ContainerStarted","Data":"266db7a9c0bcfa5ef155b35ab4e9ef87e78ba77d6daace95769021b26bed7452"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.858846 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerStarted","Data":"95dbbc2b264de16726fd5cf962dcf5b2ea23cce70901b014633e800857d2e6ac"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.879756 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerStarted","Data":"48413a96dbd55c4eac4e40c0f894a4fe721bcad741ce542a08321841cf77afd6"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.883921 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerStarted","Data":"54e4993f930974a911028d70d8cf0026f8c158c66de448c1c5a9f2c43316b560"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.886097 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-d78tr" podStartSLOduration=19.394127417 podStartE2EDuration="57.886076789s" podCreationTimestamp="2025-10-03 13:52:00 +0000 UTC" firstStartedPulling="2025-10-03 13:52:17.313484396 +0000 UTC m=+1306.516827813" lastFinishedPulling="2025-10-03 13:52:55.805433768 +0000 UTC m=+1345.008777185" observedRunningTime="2025-10-03 13:52:57.859723539 +0000 UTC m=+1347.063066966" watchObservedRunningTime="2025-10-03 13:52:57.886076789 +0000 UTC m=+1347.089420206" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.891854 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-54b8ffdcf-7hs45" podStartSLOduration=4.005622308 podStartE2EDuration="12.891838889s" podCreationTimestamp="2025-10-03 13:52:45 +0000 UTC" firstStartedPulling="2025-10-03 13:52:46.890577051 +0000 UTC m=+1336.093920468" lastFinishedPulling="2025-10-03 13:52:55.776793632 +0000 UTC m=+1344.980137049" observedRunningTime="2025-10-03 13:52:57.876424865 +0000 UTC m=+1347.079768282" watchObservedRunningTime="2025-10-03 13:52:57.891838889 +0000 UTC m=+1347.095182306" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.903034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerStarted","Data":"06d4a6beaf8c7baed60c86f5ef66a97b9dcd29bdf45ef077eba2ec59e7cf24e6"} Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.903075 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.903093 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.906605 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" podStartSLOduration=3.93280182 podStartE2EDuration="12.906589146s" podCreationTimestamp="2025-10-03 13:52:45 +0000 UTC" firstStartedPulling="2025-10-03 13:52:46.620646261 +0000 UTC m=+1335.823989678" lastFinishedPulling="2025-10-03 13:52:55.594433587 +0000 UTC m=+1344.797777004" observedRunningTime="2025-10-03 13:52:57.903521082 +0000 UTC m=+1347.106864499" watchObservedRunningTime="2025-10-03 13:52:57.906589146 +0000 UTC m=+1347.109932563" Oct 03 13:52:57 crc kubenswrapper[4959]: I1003 13:52:57.955220 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-677d776894-ljrrj" podStartSLOduration=5.955181506 podStartE2EDuration="5.955181506s" podCreationTimestamp="2025-10-03 13:52:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:57.940651823 +0000 UTC m=+1347.143995240" watchObservedRunningTime="2025-10-03 13:52:57.955181506 +0000 UTC m=+1347.158524923" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.435887 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.817768 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.914155 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerStarted","Data":"ef5bb5fa68e1b0f91f0e28923d536c8b9b0f93729fbaa329dbbb9e54e1c07088"} Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.914334 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-log" containerID="cri-o://54e4993f930974a911028d70d8cf0026f8c158c66de448c1c5a9f2c43316b560" gracePeriod=30 Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.914738 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-httpd" containerID="cri-o://ef5bb5fa68e1b0f91f0e28923d536c8b9b0f93729fbaa329dbbb9e54e1c07088" gracePeriod=30 Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.924658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerStarted","Data":"5305c7f6000e96202d843fff5e9231beafd80deb96b4784806f137e02c18a363"} Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.924806 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.927607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerStarted","Data":"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a"} Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.927627 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-log" containerID="cri-o://7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" gracePeriod=30 Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.927765 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-httpd" containerID="cri-o://d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" gracePeriod=30 Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.934394 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.951087 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=12.951069112999999 podStartE2EDuration="12.951069113s" podCreationTimestamp="2025-10-03 13:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:58.943586581 +0000 UTC m=+1348.146930008" watchObservedRunningTime="2025-10-03 13:52:58.951069113 +0000 UTC m=+1348.154412530" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.978759 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.379546709 podStartE2EDuration="14.978735224s" podCreationTimestamp="2025-10-03 13:52:44 +0000 UTC" firstStartedPulling="2025-10-03 13:52:45.587869298 +0000 UTC m=+1334.791212715" lastFinishedPulling="2025-10-03 13:52:58.187057813 +0000 UTC m=+1347.390401230" observedRunningTime="2025-10-03 13:52:58.972094633 +0000 UTC m=+1348.175438050" watchObservedRunningTime="2025-10-03 13:52:58.978735224 +0000 UTC m=+1348.182078641" Oct 03 13:52:58 crc kubenswrapper[4959]: I1003 13:52:58.996870 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.996851054 podStartE2EDuration="12.996851054s" podCreationTimestamp="2025-10-03 13:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:52:58.992913698 +0000 UTC m=+1348.196257135" watchObservedRunningTime="2025-10-03 13:52:58.996851054 +0000 UTC m=+1348.200194461" Oct 03 13:52:59 crc kubenswrapper[4959]: I1003 13:52:59.556468 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:52:59 crc kubenswrapper[4959]: I1003 13:52:59.923064 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.043510 4959 generic.go:334] "Generic (PLEG): container finished" podID="440fa439-935e-4cba-8da7-89b2f6a797ea" containerID="cfae109f6f3042bf10c7d8abaf5600ce7fac83b01ce2f4be11acd687b892aff1" exitCode=0 Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.043707 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tdxpk" event={"ID":"440fa439-935e-4cba-8da7-89b2f6a797ea","Type":"ContainerDied","Data":"cfae109f6f3042bf10c7d8abaf5600ce7fac83b01ce2f4be11acd687b892aff1"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.093978 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094070 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094289 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094338 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094374 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg7cb\" (UniqueName: \"kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094497 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.094548 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts\") pod \"429beb30-1784-4027-a0f0-230b9e864dcf\" (UID: \"429beb30-1784-4027-a0f0-230b9e864dcf\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.102335 4959 generic.go:334] "Generic (PLEG): container finished" podID="0c939246-dadd-4991-b6cd-a4767caa8931" containerID="ef5bb5fa68e1b0f91f0e28923d536c8b9b0f93729fbaa329dbbb9e54e1c07088" exitCode=0 Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.102979 4959 generic.go:334] "Generic (PLEG): container finished" podID="0c939246-dadd-4991-b6cd-a4767caa8931" containerID="54e4993f930974a911028d70d8cf0026f8c158c66de448c1c5a9f2c43316b560" exitCode=143 Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.103137 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerDied","Data":"ef5bb5fa68e1b0f91f0e28923d536c8b9b0f93729fbaa329dbbb9e54e1c07088"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.103585 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerDied","Data":"54e4993f930974a911028d70d8cf0026f8c158c66de448c1c5a9f2c43316b560"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.107343 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.107623 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts" (OuterVolumeSpecName: "scripts") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.107825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs" (OuterVolumeSpecName: "logs") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.109570 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.114744 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb" (OuterVolumeSpecName: "kube-api-access-mg7cb") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "kube-api-access-mg7cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.120234 4959 generic.go:334] "Generic (PLEG): container finished" podID="429beb30-1784-4027-a0f0-230b9e864dcf" containerID="d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" exitCode=143 Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.120261 4959 generic.go:334] "Generic (PLEG): container finished" podID="429beb30-1784-4027-a0f0-230b9e864dcf" containerID="7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" exitCode=143 Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.121110 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.121507 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerDied","Data":"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.121531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerDied","Data":"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.121599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"429beb30-1784-4027-a0f0-230b9e864dcf","Type":"ContainerDied","Data":"21f62d26ac6df23d7b61c9030bf20759f98cdcfcf15a1c24b12f35cfbb26f4ee"} Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.121614 4959 scope.go:117] "RemoveContainer" containerID="d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.150359 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200341 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200372 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200393 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200404 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200413 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg7cb\" (UniqueName: \"kubernetes.io/projected/429beb30-1784-4027-a0f0-230b9e864dcf-kube-api-access-mg7cb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.200423 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/429beb30-1784-4027-a0f0-230b9e864dcf-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.208676 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data" (OuterVolumeSpecName: "config-data") pod "429beb30-1784-4027-a0f0-230b9e864dcf" (UID: "429beb30-1784-4027-a0f0-230b9e864dcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.234259 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.302173 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/429beb30-1784-4027-a0f0-230b9e864dcf-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.302227 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.361856 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.377373 4959 scope.go:117] "RemoveContainer" containerID="7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.485692 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.492407 4959 scope.go:117] "RemoveContainer" containerID="d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.494087 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a\": container with ID starting with d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a not found: ID does not exist" containerID="d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.494135 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a"} err="failed to get container status \"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a\": rpc error: code = NotFound desc = could not find container \"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a\": container with ID starting with d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a not found: ID does not exist" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.494167 4959 scope.go:117] "RemoveContainer" containerID="7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.495281 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.498569 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab\": container with ID starting with 7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab not found: ID does not exist" containerID="7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.498698 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab"} err="failed to get container status \"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab\": rpc error: code = NotFound desc = could not find container \"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab\": container with ID starting with 7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab not found: ID does not exist" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.498787 4959 scope.go:117] "RemoveContainer" containerID="d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.506832 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w87dm\" (UniqueName: \"kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.506915 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.506938 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.507026 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.507080 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.507323 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.507351 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data\") pod \"0c939246-dadd-4991-b6cd-a4767caa8931\" (UID: \"0c939246-dadd-4991-b6cd-a4767caa8931\") " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.508110 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.509458 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs" (OuterVolumeSpecName: "logs") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.515065 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm" (OuterVolumeSpecName: "kube-api-access-w87dm") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "kube-api-access-w87dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.517652 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.517989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.518003 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.518016 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.518022 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.518049 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.518056 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: E1003 13:53:00.518066 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.518073 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.520363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts" (OuterVolumeSpecName: "scripts") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.520410 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a"} err="failed to get container status \"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a\": rpc error: code = NotFound desc = could not find container \"d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a\": container with ID starting with d90d82469e6aa1852c38c173d1966c951eaa45942225304dbcde9ec30d16a08a not found: ID does not exist" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.520611 4959 scope.go:117] "RemoveContainer" containerID="7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.521672 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab"} err="failed to get container status \"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab\": rpc error: code = NotFound desc = could not find container \"7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab\": container with ID starting with 7a00776cfabcbbd695e4342fd9f5586ece76d1e38aa0be2e55b22f63c01434ab not found: ID does not exist" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.524483 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.524540 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.524572 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-log" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.524585 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" containerName="glance-httpd" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.541572 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.550768 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.554957 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.555750 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.564321 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.590696 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.600589 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data" (OuterVolumeSpecName: "config-data") pod "0c939246-dadd-4991-b6cd-a4767caa8931" (UID: "0c939246-dadd-4991-b6cd-a4767caa8931"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609600 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609672 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609687 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w87dm\" (UniqueName: \"kubernetes.io/projected/0c939246-dadd-4991-b6cd-a4767caa8931-kube-api-access-w87dm\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609721 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609731 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609740 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c939246-dadd-4991-b6cd-a4767caa8931-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.609750 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c939246-dadd-4991-b6cd-a4767caa8931-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.639828 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711313 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711432 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711697 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711782 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711847 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mtnm\" (UniqueName: \"kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.711960 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.712035 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814054 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814105 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814170 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814218 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mtnm\" (UniqueName: \"kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814244 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814740 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.814874 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.815528 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.818942 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.818961 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.820535 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.830477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.833636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mtnm\" (UniqueName: \"kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.848319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:00 crc kubenswrapper[4959]: I1003 13:53:00.891016 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.139620 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c939246-dadd-4991-b6cd-a4767caa8931","Type":"ContainerDied","Data":"e6c648a2da50214bcfce64fe3f34cc572ea4c8df28649f9ef5c9193e8e513896"} Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.139914 4959 scope.go:117] "RemoveContainer" containerID="ef5bb5fa68e1b0f91f0e28923d536c8b9b0f93729fbaa329dbbb9e54e1c07088" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.140052 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.208254 4959 scope.go:117] "RemoveContainer" containerID="54e4993f930974a911028d70d8cf0026f8c158c66de448c1c5a9f2c43316b560" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.213566 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.256943 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.265833 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.267438 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.274170 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.274398 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.275254 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431139 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431243 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431302 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431379 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6cvb\" (UniqueName: \"kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431483 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.431509 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.512730 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:01 crc kubenswrapper[4959]: W1003 13:53:01.520002 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cea2d9b_538b_4333_a81f_c59792924b06.slice/crio-64af77666280070d4b3c2e94d1891640387ed89f27033f521e471b6905e84a84 WatchSource:0}: Error finding container 64af77666280070d4b3c2e94d1891640387ed89f27033f521e471b6905e84a84: Status 404 returned error can't find the container with id 64af77666280070d4b3c2e94d1891640387ed89f27033f521e471b6905e84a84 Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.532923 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.534264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.534669 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.534995 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.535285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.535527 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6cvb\" (UniqueName: \"kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.535665 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.535870 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.536060 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.535522 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.536997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.540396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.540560 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.543210 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.545396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.555024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6cvb\" (UniqueName: \"kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.577275 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.589566 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.598290 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.621501 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.710506 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c939246-dadd-4991-b6cd-a4767caa8931" path="/var/lib/kubelet/pods/0c939246-dadd-4991-b6cd-a4767caa8931/volumes" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.711580 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="429beb30-1784-4027-a0f0-230b9e864dcf" path="/var/lib/kubelet/pods/429beb30-1784-4027-a0f0-230b9e864dcf/volumes" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.727641 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.727966 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="dnsmasq-dns" containerID="cri-o://c57277a9f005fbbf82e51c493692d3d1b8c379c4ff89dcced7da50fa7ecb7805" gracePeriod=10 Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.744900 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle\") pod \"440fa439-935e-4cba-8da7-89b2f6a797ea\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.745325 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config\") pod \"440fa439-935e-4cba-8da7-89b2f6a797ea\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.746085 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kldk9\" (UniqueName: \"kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9\") pod \"440fa439-935e-4cba-8da7-89b2f6a797ea\" (UID: \"440fa439-935e-4cba-8da7-89b2f6a797ea\") " Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.752010 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9" (OuterVolumeSpecName: "kube-api-access-kldk9") pod "440fa439-935e-4cba-8da7-89b2f6a797ea" (UID: "440fa439-935e-4cba-8da7-89b2f6a797ea"). InnerVolumeSpecName "kube-api-access-kldk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.772232 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config" (OuterVolumeSpecName: "config") pod "440fa439-935e-4cba-8da7-89b2f6a797ea" (UID: "440fa439-935e-4cba-8da7-89b2f6a797ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.777517 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "440fa439-935e-4cba-8da7-89b2f6a797ea" (UID: "440fa439-935e-4cba-8da7-89b2f6a797ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.848652 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kldk9\" (UniqueName: \"kubernetes.io/projected/440fa439-935e-4cba-8da7-89b2f6a797ea-kube-api-access-kldk9\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.849028 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:01 crc kubenswrapper[4959]: I1003 13:53:01.849042 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/440fa439-935e-4cba-8da7-89b2f6a797ea-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.169768 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerStarted","Data":"64af77666280070d4b3c2e94d1891640387ed89f27033f521e471b6905e84a84"} Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.175751 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tdxpk" event={"ID":"440fa439-935e-4cba-8da7-89b2f6a797ea","Type":"ContainerDied","Data":"a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9"} Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.175789 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a63844b6c8ddc46d55bf580108090ff0c823a2f8892b5a5370dad5f7f625ccc9" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.175853 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tdxpk" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.224662 4959 generic.go:334] "Generic (PLEG): container finished" podID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerID="c57277a9f005fbbf82e51c493692d3d1b8c379c4ff89dcced7da50fa7ecb7805" exitCode=0 Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.224737 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerDied","Data":"c57277a9f005fbbf82e51c493692d3d1b8c379c4ff89dcced7da50fa7ecb7805"} Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.268515 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:02 crc kubenswrapper[4959]: E1003 13:53:02.269055 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440fa439-935e-4cba-8da7-89b2f6a797ea" containerName="neutron-db-sync" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.269234 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="440fa439-935e-4cba-8da7-89b2f6a797ea" containerName="neutron-db-sync" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.269591 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="440fa439-935e-4cba-8da7-89b2f6a797ea" containerName="neutron-db-sync" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.285080 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.292787 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jcjmw" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.293859 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.294358 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.298052 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.304081 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.314936 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.329987 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.340479 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444085 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkpkf\" (UniqueName: \"kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444476 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdt5h\" (UniqueName: \"kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444521 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444553 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444687 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444742 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.444876 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.445015 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.500152 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546401 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546475 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546509 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkpkf\" (UniqueName: \"kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546629 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdt5h\" (UniqueName: \"kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546664 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546737 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.546779 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.547707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.548970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.549685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.555966 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.560319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.561596 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.565957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.573149 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.578653 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.586024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkpkf\" (UniqueName: \"kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf\") pod \"dnsmasq-dns-75c8ddd69c-xm82z\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.596013 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdt5h\" (UniqueName: \"kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h\") pod \"neutron-596475bc7b-44mtq\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.607922 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.643123 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.647928 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.647996 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.648140 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.651935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.652742 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.652897 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4528s\" (UniqueName: \"kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s\") pod \"d501eb7e-a318-4ecd-8fad-109155ba0763\" (UID: \"d501eb7e-a318-4ecd-8fad-109155ba0763\") " Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.667660 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s" (OuterVolumeSpecName: "kube-api-access-4528s") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "kube-api-access-4528s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.727084 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.742375 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.757174 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.765394 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.765421 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4528s\" (UniqueName: \"kubernetes.io/projected/d501eb7e-a318-4ecd-8fad-109155ba0763-kube-api-access-4528s\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.765430 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.771505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.790957 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config" (OuterVolumeSpecName: "config") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.791958 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d501eb7e-a318-4ecd-8fad-109155ba0763" (UID: "d501eb7e-a318-4ecd-8fad-109155ba0763"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.879457 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.879478 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:02 crc kubenswrapper[4959]: I1003 13:53:02.879488 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d501eb7e-a318-4ecd-8fad-109155ba0763-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.158245 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.269325 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" event={"ID":"ccca3d9a-e8e7-4699-8713-21d6ce4445c1","Type":"ContainerStarted","Data":"389a8010bc94b8a851e03b2781d4f8c794eec965742f3e8c5e2acada0c988934"} Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.281221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerStarted","Data":"895ba025e112e38efeebd5087798538e71bd491fd6f4aa9c4ad1eec72c06f4a8"} Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.311471 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" event={"ID":"d501eb7e-a318-4ecd-8fad-109155ba0763","Type":"ContainerDied","Data":"b3730a68d1e6c5a0d664358618dc86b6184f00b22ac8d1b70d5e5a8acf83482b"} Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.311529 4959 scope.go:117] "RemoveContainer" containerID="c57277a9f005fbbf82e51c493692d3d1b8c379c4ff89dcced7da50fa7ecb7805" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.311664 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-jhcxt" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.360588 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.385544 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.406122 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-jhcxt"] Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.444300 4959 scope.go:117] "RemoveContainer" containerID="2005a597211cd6d3b56104e48debb5f573011008cd099ef931a10f3a11e36164" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.606345 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:53:03 crc kubenswrapper[4959]: E1003 13:53:03.606713 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="init" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.606728 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="init" Oct 03 13:53:03 crc kubenswrapper[4959]: E1003 13:53:03.606750 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="dnsmasq-dns" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.606756 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="dnsmasq-dns" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.606935 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" containerName="dnsmasq-dns" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.607929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.614478 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.614720 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.614869 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.628932 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.699683 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.699761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.699803 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.699841 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.700087 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.700146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km9l4\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.700332 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.700463 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.704296 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d501eb7e-a318-4ecd-8fad-109155ba0763" path="/var/lib/kubelet/pods/d501eb7e-a318-4ecd-8fad-109155ba0763/volumes" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804126 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804280 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804304 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804355 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804398 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.804430 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km9l4\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.806155 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.821761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.833341 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.837913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.838648 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.840899 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.842892 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km9l4\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.849299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle\") pod \"swift-proxy-5b84954875-j5bdp\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:03 crc kubenswrapper[4959]: I1003 13:53:03.952096 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.384809 4959 generic.go:334] "Generic (PLEG): container finished" podID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerID="812e6e0224c1b813604ba1dc36cb6511db46d98c9630a312f4db5a4e6c19e769" exitCode=0 Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.385177 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" event={"ID":"ccca3d9a-e8e7-4699-8713-21d6ce4445c1","Type":"ContainerDied","Data":"812e6e0224c1b813604ba1dc36cb6511db46d98c9630a312f4db5a4e6c19e769"} Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.392472 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerStarted","Data":"6f65f901047f072024008154bcb5706dd05c5c1c069a77f6cce157da11e35684"} Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.406692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerStarted","Data":"8c164c5247a2631e33c76e1ec7e46302a319f9bf2b75393b0785b327b5b1ed77"} Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.422721 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerStarted","Data":"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc"} Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.422763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerStarted","Data":"2744e49cf24e053f183c0f3a9e6ce49a3599757356b98393c127fcd8c8b972d0"} Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.525320 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.526822 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.530549 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.537306 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.554882 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.644701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645059 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645091 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d99lt\" (UniqueName: \"kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645540 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.645746 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747314 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747361 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747393 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747428 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747450 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.747508 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d99lt\" (UniqueName: \"kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.754980 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.755872 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.757938 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.766972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.779938 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d99lt\" (UniqueName: \"kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.780597 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.786272 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs\") pod \"neutron-76dd7cdf5-g59pr\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.833443 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:53:04 crc kubenswrapper[4959]: W1003 13:53:04.871768 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf325dd29_3b0d_4188_9ac2_e681d56e881f.slice/crio-d0cfb94ed897d811d8aa2d0ef9731b2fa0437d894fa179674d671b3d8965b0cc WatchSource:0}: Error finding container d0cfb94ed897d811d8aa2d0ef9731b2fa0437d894fa179674d671b3d8965b0cc: Status 404 returned error can't find the container with id d0cfb94ed897d811d8aa2d0ef9731b2fa0437d894fa179674d671b3d8965b0cc Oct 03 13:53:04 crc kubenswrapper[4959]: I1003 13:53:04.902902 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.042587 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.440386 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerStarted","Data":"99ae8be8736d1264a4e5a30b3acec9a9ccd988ea8c5766acc11621fb435ee82c"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.458531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerStarted","Data":"87681968bf3b43d810f5c797f220cf83f293f5af560789c5caf22b22226ad5c3"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.464641 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerStarted","Data":"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.465437 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.470644 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.470629723 podStartE2EDuration="5.470629723s" podCreationTimestamp="2025-10-03 13:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:05.469465664 +0000 UTC m=+1354.672809081" watchObservedRunningTime="2025-10-03 13:53:05.470629723 +0000 UTC m=+1354.673973140" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.476428 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerStarted","Data":"802921468f08beb7a7faecf994bffea767c35ceda75b15525c02a4e2c871e13a"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.476478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerStarted","Data":"d0cfb94ed897d811d8aa2d0ef9731b2fa0437d894fa179674d671b3d8965b0cc"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.483241 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" event={"ID":"ccca3d9a-e8e7-4699-8713-21d6ce4445c1","Type":"ContainerStarted","Data":"9420f8084d458dabaa75af29ba5a0a48d6ea4846db169c617c705cda323e00ef"} Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.484053 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.513810 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-596475bc7b-44mtq" podStartSLOduration=3.513787939 podStartE2EDuration="3.513787939s" podCreationTimestamp="2025-10-03 13:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:05.502494245 +0000 UTC m=+1354.705837662" watchObservedRunningTime="2025-10-03 13:53:05.513787939 +0000 UTC m=+1354.717131356" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.547352 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" podStartSLOduration=3.547332243 podStartE2EDuration="3.547332243s" podCreationTimestamp="2025-10-03 13:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:05.545929309 +0000 UTC m=+1354.749272726" watchObservedRunningTime="2025-10-03 13:53:05.547332243 +0000 UTC m=+1354.750675660" Oct 03 13:53:05 crc kubenswrapper[4959]: I1003 13:53:05.648296 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:53:05 crc kubenswrapper[4959]: W1003 13:53:05.649332 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84716cd9_f863_4713_bc0e_15e719ac29d2.slice/crio-e98eb48a73086f65135a709ca43532fe7c2309a0b2ab502d99924bf749bfb87e WatchSource:0}: Error finding container e98eb48a73086f65135a709ca43532fe7c2309a0b2ab502d99924bf749bfb87e: Status 404 returned error can't find the container with id e98eb48a73086f65135a709ca43532fe7c2309a0b2ab502d99924bf749bfb87e Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.398031 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.464793 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.465358 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" containerID="cri-o://be3ac4d3f6020be7c862df77e04e33b60d571c907f1325c96ff42d1be4628d4f" gracePeriod=30 Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.465667 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" containerID="cri-o://004a0ee4f59f60160da67b71b5e1962f4e4473c3ff1b2699c4ef12fcaa6e500d" gracePeriod=30 Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.502355 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerStarted","Data":"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808"} Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.502398 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerStarted","Data":"e98eb48a73086f65135a709ca43532fe7c2309a0b2ab502d99924bf749bfb87e"} Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.504144 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerStarted","Data":"e51713b9b874f135af735bd09481738fbf57ba1d4691649ef78b87b0fca43cd1"} Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.505156 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.505209 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.505392 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": EOF" Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.536667 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5b84954875-j5bdp" podStartSLOduration=3.536645521 podStartE2EDuration="3.536645521s" podCreationTimestamp="2025-10-03 13:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:06.528591786 +0000 UTC m=+1355.731935203" watchObservedRunningTime="2025-10-03 13:53:06.536645521 +0000 UTC m=+1355.739988938" Oct 03 13:53:06 crc kubenswrapper[4959]: I1003 13:53:06.581096 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.5810735000000005 podStartE2EDuration="5.5810735s" podCreationTimestamp="2025-10-03 13:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:06.564833555 +0000 UTC m=+1355.768176972" watchObservedRunningTime="2025-10-03 13:53:06.5810735 +0000 UTC m=+1355.784416917" Oct 03 13:53:07 crc kubenswrapper[4959]: I1003 13:53:07.517323 4959 generic.go:334] "Generic (PLEG): container finished" podID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerID="be3ac4d3f6020be7c862df77e04e33b60d571c907f1325c96ff42d1be4628d4f" exitCode=143 Oct 03 13:53:07 crc kubenswrapper[4959]: I1003 13:53:07.517392 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerDied","Data":"be3ac4d3f6020be7c862df77e04e33b60d571c907f1325c96ff42d1be4628d4f"} Oct 03 13:53:07 crc kubenswrapper[4959]: I1003 13:53:07.519857 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerStarted","Data":"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6"} Oct 03 13:53:07 crc kubenswrapper[4959]: I1003 13:53:07.540858 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76dd7cdf5-g59pr" podStartSLOduration=3.54083541 podStartE2EDuration="3.54083541s" podCreationTimestamp="2025-10-03 13:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:07.53715288 +0000 UTC m=+1356.740496317" watchObservedRunningTime="2025-10-03 13:53:07.54083541 +0000 UTC m=+1356.744178827" Oct 03 13:53:08 crc kubenswrapper[4959]: I1003 13:53:08.529450 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.464732 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.465907 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-central-agent" containerID="cri-o://d1a6cf72b4294e6230e6e22e8ebf1e91d9bdb34958530232770c3565f560f85b" gracePeriod=30 Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.466077 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" containerID="cri-o://5305c7f6000e96202d843fff5e9231beafd80deb96b4784806f137e02c18a363" gracePeriod=30 Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.466136 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="sg-core" containerID="cri-o://282699a695dc3c74712766e4667ffd36463c08f2553c45fa465d271ca4ae5b72" gracePeriod=30 Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.466179 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-notification-agent" containerID="cri-o://116ab8d592eea1dea5fa83f80f343889469bd35b63bf2d67c31a6881bdda29c1" gracePeriod=30 Oct 03 13:53:09 crc kubenswrapper[4959]: I1003 13:53:09.476600 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.142:3000/\": EOF" Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.555824 4959 generic.go:334] "Generic (PLEG): container finished" podID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerID="004a0ee4f59f60160da67b71b5e1962f4e4473c3ff1b2699c4ef12fcaa6e500d" exitCode=0 Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.555924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerDied","Data":"004a0ee4f59f60160da67b71b5e1962f4e4473c3ff1b2699c4ef12fcaa6e500d"} Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559323 4959 generic.go:334] "Generic (PLEG): container finished" podID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerID="5305c7f6000e96202d843fff5e9231beafd80deb96b4784806f137e02c18a363" exitCode=0 Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559368 4959 generic.go:334] "Generic (PLEG): container finished" podID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerID="282699a695dc3c74712766e4667ffd36463c08f2553c45fa465d271ca4ae5b72" exitCode=2 Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559385 4959 generic.go:334] "Generic (PLEG): container finished" podID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerID="116ab8d592eea1dea5fa83f80f343889469bd35b63bf2d67c31a6881bdda29c1" exitCode=0 Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559398 4959 generic.go:334] "Generic (PLEG): container finished" podID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerID="d1a6cf72b4294e6230e6e22e8ebf1e91d9bdb34958530232770c3565f560f85b" exitCode=0 Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerDied","Data":"5305c7f6000e96202d843fff5e9231beafd80deb96b4784806f137e02c18a363"} Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559487 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerDied","Data":"282699a695dc3c74712766e4667ffd36463c08f2553c45fa465d271ca4ae5b72"} Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559514 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerDied","Data":"116ab8d592eea1dea5fa83f80f343889469bd35b63bf2d67c31a6881bdda29c1"} Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.559534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerDied","Data":"d1a6cf72b4294e6230e6e22e8ebf1e91d9bdb34958530232770c3565f560f85b"} Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.891911 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.892590 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.943300 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:10 crc kubenswrapper[4959]: I1003 13:53:10.949843 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.569003 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.569052 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.600263 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.600344 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.650915 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:53:11 crc kubenswrapper[4959]: I1003 13:53:11.654899 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:53:12 crc kubenswrapper[4959]: I1003 13:53:12.580399 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:53:12 crc kubenswrapper[4959]: I1003 13:53:12.580733 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:53:12 crc kubenswrapper[4959]: I1003 13:53:12.758391 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:12 crc kubenswrapper[4959]: I1003 13:53:12.844615 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:53:12 crc kubenswrapper[4959]: I1003 13:53:12.848499 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="dnsmasq-dns" containerID="cri-o://9e99639ccaed7b57e2b9ffe2295046f5e3e386fac427e02a3d1ce7f2f3732062" gracePeriod=10 Oct 03 13:53:13 crc kubenswrapper[4959]: I1003 13:53:13.594004 4959 generic.go:334] "Generic (PLEG): container finished" podID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerID="9e99639ccaed7b57e2b9ffe2295046f5e3e386fac427e02a3d1ce7f2f3732062" exitCode=0 Oct 03 13:53:13 crc kubenswrapper[4959]: I1003 13:53:13.594245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" event={"ID":"e5e4fb74-f494-4b58-8ea2-1af811580474","Type":"ContainerDied","Data":"9e99639ccaed7b57e2b9ffe2295046f5e3e386fac427e02a3d1ce7f2f3732062"} Oct 03 13:53:13 crc kubenswrapper[4959]: I1003 13:53:13.961305 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:13 crc kubenswrapper[4959]: I1003 13:53:13.967562 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:53:14 crc kubenswrapper[4959]: I1003 13:53:14.627967 4959 generic.go:334] "Generic (PLEG): container finished" podID="1433da4c-d3b1-48ea-b62e-e70af51671e5" containerID="266db7a9c0bcfa5ef155b35ab4e9ef87e78ba77d6daace95769021b26bed7452" exitCode=0 Oct 03 13:53:14 crc kubenswrapper[4959]: I1003 13:53:14.628068 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d78tr" event={"ID":"1433da4c-d3b1-48ea-b62e-e70af51671e5","Type":"ContainerDied","Data":"266db7a9c0bcfa5ef155b35ab4e9ef87e78ba77d6daace95769021b26bed7452"} Oct 03 13:53:14 crc kubenswrapper[4959]: I1003 13:53:14.756037 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:14 crc kubenswrapper[4959]: I1003 13:53:14.756140 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:53:14 crc kubenswrapper[4959]: I1003 13:53:14.789985 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:15 crc kubenswrapper[4959]: I1003 13:53:15.068800 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:53:15 crc kubenswrapper[4959]: I1003 13:53:15.069134 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:53:15 crc kubenswrapper[4959]: I1003 13:53:15.542824 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:53:16 crc kubenswrapper[4959]: I1003 13:53:16.493874 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:53:16 crc kubenswrapper[4959]: I1003 13:53:16.493910 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:53:16 crc kubenswrapper[4959]: I1003 13:53:16.603903 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: connect: connection refused" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.549772 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.556750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.563752 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d78tr" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.610754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs\") pod \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.610808 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle\") pod \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.610910 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom\") pod \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.610965 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sdzk\" (UniqueName: \"kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk\") pod \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.611110 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data\") pod \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\" (UID: \"b074b3a0-c821-49c6-bbef-a95e02e45fb2\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.612735 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs" (OuterVolumeSpecName: "logs") pod "b074b3a0-c821-49c6-bbef-a95e02e45fb2" (UID: "b074b3a0-c821-49c6-bbef-a95e02e45fb2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.625266 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk" (OuterVolumeSpecName: "kube-api-access-6sdzk") pod "b074b3a0-c821-49c6-bbef-a95e02e45fb2" (UID: "b074b3a0-c821-49c6-bbef-a95e02e45fb2"). InnerVolumeSpecName "kube-api-access-6sdzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.628414 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b074b3a0-c821-49c6-bbef-a95e02e45fb2" (UID: "b074b3a0-c821-49c6-bbef-a95e02e45fb2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.678292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b074b3a0-c821-49c6-bbef-a95e02e45fb2" (UID: "b074b3a0-c821-49c6-bbef-a95e02e45fb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.680406 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-d78tr" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.681337 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-d78tr" event={"ID":"1433da4c-d3b1-48ea-b62e-e70af51671e5","Type":"ContainerDied","Data":"1610d5359d738f07bdb8db33ceb61e11ca2a7012f35f93fdf08636df4548a228"} Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.681375 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1610d5359d738f07bdb8db33ceb61e11ca2a7012f35f93fdf08636df4548a228" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.698838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"779685e0-3882-4dbe-ad13-df648b89d1a0","Type":"ContainerDied","Data":"78a375752b390c039898849196854c58e23d561421eca4e1b6131eeb828ce19d"} Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.698974 4959 scope.go:117] "RemoveContainer" containerID="5305c7f6000e96202d843fff5e9231beafd80deb96b4784806f137e02c18a363" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.699083 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.708006 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-684fb7ddcd-zvzm9" event={"ID":"b074b3a0-c821-49c6-bbef-a95e02e45fb2","Type":"ContainerDied","Data":"f98018770d73ca1a06941c0b77f8ce0fd0bdcaf2f9421172d872f62b005e060d"} Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.708115 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-684fb7ddcd-zvzm9" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712528 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712676 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712753 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712800 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712825 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s95b\" (UniqueName: \"kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712859 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712897 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712960 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.712990 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713014 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c266\" (UniqueName: \"kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266\") pod \"1433da4c-d3b1-48ea-b62e-e70af51671e5\" (UID: \"1433da4c-d3b1-48ea-b62e-e70af51671e5\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713039 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd\") pod \"779685e0-3882-4dbe-ad13-df648b89d1a0\" (UID: \"779685e0-3882-4dbe-ad13-df648b89d1a0\") " Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713580 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713610 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sdzk\" (UniqueName: \"kubernetes.io/projected/b074b3a0-c821-49c6-bbef-a95e02e45fb2-kube-api-access-6sdzk\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713625 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b074b3a0-c821-49c6-bbef-a95e02e45fb2-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.713636 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.714154 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.714241 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.721157 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts" (OuterVolumeSpecName: "scripts") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.723440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b" (OuterVolumeSpecName: "kube-api-access-7s95b") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "kube-api-access-7s95b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.723760 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.727359 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.727491 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts" (OuterVolumeSpecName: "scripts") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.727916 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266" (OuterVolumeSpecName: "kube-api-access-2c266") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "kube-api-access-2c266". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.771686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data" (OuterVolumeSpecName: "config-data") pod "b074b3a0-c821-49c6-bbef-a95e02e45fb2" (UID: "b074b3a0-c821-49c6-bbef-a95e02e45fb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.776503 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.779097 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.793347 4959 scope.go:117] "RemoveContainer" containerID="282699a695dc3c74712766e4667ffd36463c08f2553c45fa465d271ca4ae5b72" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815722 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b074b3a0-c821-49c6-bbef-a95e02e45fb2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815747 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1433da4c-d3b1-48ea-b62e-e70af51671e5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815757 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815766 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815774 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815783 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s95b\" (UniqueName: \"kubernetes.io/projected/779685e0-3882-4dbe-ad13-df648b89d1a0-kube-api-access-7s95b\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815791 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815800 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815808 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815816 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c266\" (UniqueName: \"kubernetes.io/projected/1433da4c-d3b1-48ea-b62e-e70af51671e5-kube-api-access-2c266\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.815823 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/779685e0-3882-4dbe-ad13-df648b89d1a0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.824928 4959 scope.go:117] "RemoveContainer" containerID="116ab8d592eea1dea5fa83f80f343889469bd35b63bf2d67c31a6881bdda29c1" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.853616 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.857453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data" (OuterVolumeSpecName: "config-data") pod "1433da4c-d3b1-48ea-b62e-e70af51671e5" (UID: "1433da4c-d3b1-48ea-b62e-e70af51671e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.869499 4959 scope.go:117] "RemoveContainer" containerID="d1a6cf72b4294e6230e6e22e8ebf1e91d9bdb34958530232770c3565f560f85b" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.906314 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data" (OuterVolumeSpecName: "config-data") pod "779685e0-3882-4dbe-ad13-df648b89d1a0" (UID: "779685e0-3882-4dbe-ad13-df648b89d1a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.917826 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.917853 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1433da4c-d3b1-48ea-b62e-e70af51671e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.917863 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/779685e0-3882-4dbe-ad13-df648b89d1a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.931457 4959 scope.go:117] "RemoveContainer" containerID="004a0ee4f59f60160da67b71b5e1962f4e4473c3ff1b2699c4ef12fcaa6e500d" Oct 03 13:53:18 crc kubenswrapper[4959]: I1003 13:53:18.965924 4959 scope.go:117] "RemoveContainer" containerID="be3ac4d3f6020be7c862df77e04e33b60d571c907f1325c96ff42d1be4628d4f" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.048552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.070708 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.076362 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.080572 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.082542 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" containerName="cinder-db-sync" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.082751 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" containerName="cinder-db-sync" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.082837 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="dnsmasq-dns" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.082893 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="dnsmasq-dns" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.082951 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="sg-core" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.083072 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="sg-core" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.083148 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-central-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.083255 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-central-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.083332 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="init" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.083423 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="init" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.083503 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.083603 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.083733 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-notification-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.083793 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-notification-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.083852 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.091422 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.091547 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.091607 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.091975 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092584 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092668 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-notification-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092737 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="sg-core" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092806 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" containerName="dnsmasq-dns" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092876 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.092964 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="ceilometer-central-agent" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.093065 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" containerName="cinder-db-sync" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.096835 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.096921 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.101605 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.105440 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.106923 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125372 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125463 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125488 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125570 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125734 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.125825 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92rzf\" (UniqueName: \"kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf\") pod \"e5e4fb74-f494-4b58-8ea2-1af811580474\" (UID: \"e5e4fb74-f494-4b58-8ea2-1af811580474\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.135594 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf" (OuterVolumeSpecName: "kube-api-access-92rzf") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "kube-api-access-92rzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.140915 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-684fb7ddcd-zvzm9"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.189059 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.203505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config" (OuterVolumeSpecName: "config") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.208077 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.210836 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.217733 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5e4fb74-f494-4b58-8ea2-1af811580474" (UID: "e5e4fb74-f494-4b58-8ea2-1af811580474"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227304 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227389 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227471 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjc7p\" (UniqueName: \"kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227571 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227624 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227639 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227648 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227660 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227671 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92rzf\" (UniqueName: \"kubernetes.io/projected/e5e4fb74-f494-4b58-8ea2-1af811580474-kube-api-access-92rzf\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.227683 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5e4fb74-f494-4b58-8ea2-1af811580474-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.328807 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.328879 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.328944 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjc7p\" (UniqueName: \"kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.328969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.329019 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.329063 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.329093 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.329544 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.330339 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.331152 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.334133 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.334680 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.335450 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: E1003 13:53:19.335873 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data kube-api-access-bjc7p scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="d20197b2-395b-4c65-81a1-0c76fd396eb2" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.337057 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.363747 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjc7p\" (UniqueName: \"kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p\") pod \"ceilometer-0\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.698160 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" path="/var/lib/kubelet/pods/779685e0-3882-4dbe-ad13-df648b89d1a0/volumes" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.699794 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" path="/var/lib/kubelet/pods/b074b3a0-c821-49c6-bbef-a95e02e45fb2/volumes" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.717682 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.717693 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-wj4hv" event={"ID":"e5e4fb74-f494-4b58-8ea2-1af811580474","Type":"ContainerDied","Data":"c1d4f567be5ac79300618904629545474c97feab4e5c4ad38b36a213a0334a3f"} Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.717950 4959 scope.go:117] "RemoveContainer" containerID="9e99639ccaed7b57e2b9ffe2295046f5e3e386fac427e02a3d1ce7f2f3732062" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.720851 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.737753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.747476 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.758612 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-wj4hv"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.839469 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.839786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.839905 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.840062 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjc7p\" (UniqueName: \"kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.840214 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.840319 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.840390 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle\") pod \"d20197b2-395b-4c65-81a1-0c76fd396eb2\" (UID: \"d20197b2-395b-4c65-81a1-0c76fd396eb2\") " Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.841559 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.841578 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.848390 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.848418 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p" (OuterVolumeSpecName: "kube-api-access-bjc7p") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "kube-api-access-bjc7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.849078 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.852277 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts" (OuterVolumeSpecName: "scripts") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.852829 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data" (OuterVolumeSpecName: "config-data") pod "d20197b2-395b-4c65-81a1-0c76fd396eb2" (UID: "d20197b2-395b-4c65-81a1-0c76fd396eb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.900098 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.907932 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.910370 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6hgc5" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.910572 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.910613 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.910768 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.923999 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945100 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945145 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945157 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945171 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945184 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d20197b2-395b-4c65-81a1-0c76fd396eb2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945233 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d20197b2-395b-4c65-81a1-0c76fd396eb2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:19 crc kubenswrapper[4959]: I1003 13:53:19.945244 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjc7p\" (UniqueName: \"kubernetes.io/projected/d20197b2-395b-4c65-81a1-0c76fd396eb2-kube-api-access-bjc7p\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.046651 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.046706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.046744 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.046982 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.047029 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.047058 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvjbs\" (UniqueName: \"kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.058666 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.060575 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.088598 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149528 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149610 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149637 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149674 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkjqc\" (UniqueName: \"kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149733 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149758 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149788 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149857 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149930 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.149960 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvjbs\" (UniqueName: \"kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.150050 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.150091 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.150207 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.155120 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.156598 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.157488 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.158462 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.169432 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvjbs\" (UniqueName: \"kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs\") pod \"cinder-scheduler-0\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.213591 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.217482 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.227914 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.231770 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251646 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkjqc\" (UniqueName: \"kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251689 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.251714 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.252509 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.253064 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.253406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.253864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.254011 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.275752 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.294405 4959 scope.go:117] "RemoveContainer" containerID="e8e43310c8bff99fcaeaf193ddceb252a39706c5515adc44ff056f2f6182f712" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.301950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkjqc\" (UniqueName: \"kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc\") pod \"dnsmasq-dns-5784cf869f-nqkwd\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.352827 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.352894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.352931 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.352993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.353036 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.353058 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.353110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55dq\" (UniqueName: \"kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.401879 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454396 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454461 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454554 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55dq\" (UniqueName: \"kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454593 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454612 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.454631 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.455229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.455415 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.458014 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.458436 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.458651 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.462540 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.476382 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55dq\" (UniqueName: \"kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq\") pod \"cinder-api-0\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.616155 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.729634 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.778565 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.805789 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.822244 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.829706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.842335 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.842868 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.857650 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962377 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962489 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962571 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h45qr\" (UniqueName: \"kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962650 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962855 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:20 crc kubenswrapper[4959]: I1003 13:53:20.962995 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064572 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064640 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064737 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064759 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h45qr\" (UniqueName: \"kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064786 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.064834 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.065813 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.066090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.068716 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.068786 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.069702 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.070256 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.091066 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h45qr\" (UniqueName: \"kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr\") pod \"ceilometer-0\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.158269 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.495159 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.495260 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-684fb7ddcd-zvzm9" podUID="b074b3a0-c821-49c6-bbef-a95e02e45fb2" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.695930 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d20197b2-395b-4c65-81a1-0c76fd396eb2" path="/var/lib/kubelet/pods/d20197b2-395b-4c65-81a1-0c76fd396eb2/volumes" Oct 03 13:53:21 crc kubenswrapper[4959]: I1003 13:53:21.696361 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e4fb74-f494-4b58-8ea2-1af811580474" path="/var/lib/kubelet/pods/e5e4fb74-f494-4b58-8ea2-1af811580474/volumes" Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.443680 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.458158 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.574553 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:53:22 crc kubenswrapper[4959]: W1003 13:53:22.574777 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb13a531c_8bcc_46b4_b034_525a5774081d.slice/crio-9331d28cadff0d1e1ad67e4b6d795ccac93a13113a776c8dc43807710ceff864 WatchSource:0}: Error finding container 9331d28cadff0d1e1ad67e4b6d795ccac93a13113a776c8dc43807710ceff864: Status 404 returned error can't find the container with id 9331d28cadff0d1e1ad67e4b6d795ccac93a13113a776c8dc43807710ceff864 Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.623653 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:22 crc kubenswrapper[4959]: W1003 13:53:22.624051 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1cfe967_4129_4b6b_a6e9_1ec8c96a00d5.slice/crio-174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a WatchSource:0}: Error finding container 174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a: Status 404 returned error can't find the container with id 174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.744315 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.758741 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" event={"ID":"b13a531c-8bcc-46b4-b034-525a5774081d","Type":"ContainerStarted","Data":"9331d28cadff0d1e1ad67e4b6d795ccac93a13113a776c8dc43807710ceff864"} Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.761539 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a7088287-7362-42f5-987c-be9f700c50cb","Type":"ContainerStarted","Data":"bae04dd3c4a3a80e408d2aa1da0eb1a0f82c686e02752a0f01f6986ff0f0542b"} Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.764863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerStarted","Data":"2d2285819d300c3728cf7e45ef5f5afa78cd7d40ed582f6081484149703a2428"} Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.767607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerStarted","Data":"8e7a71c916d834fbba63543e377476859b078486bef96e681fea5f4f3333bd2a"} Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.772577 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerStarted","Data":"174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a"} Oct 03 13:53:22 crc kubenswrapper[4959]: I1003 13:53:22.798853 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.33721148 podStartE2EDuration="28.798828215s" podCreationTimestamp="2025-10-03 13:52:54 +0000 UTC" firstStartedPulling="2025-10-03 13:52:56.403277066 +0000 UTC m=+1345.606620483" lastFinishedPulling="2025-10-03 13:53:21.864893791 +0000 UTC m=+1371.068237218" observedRunningTime="2025-10-03 13:53:22.782532249 +0000 UTC m=+1371.985875676" watchObservedRunningTime="2025-10-03 13:53:22.798828215 +0000 UTC m=+1372.002171632" Oct 03 13:53:23 crc kubenswrapper[4959]: I1003 13:53:23.788825 4959 generic.go:334] "Generic (PLEG): container finished" podID="b13a531c-8bcc-46b4-b034-525a5774081d" containerID="aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83" exitCode=0 Oct 03 13:53:23 crc kubenswrapper[4959]: I1003 13:53:23.789107 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" event={"ID":"b13a531c-8bcc-46b4-b034-525a5774081d","Type":"ContainerDied","Data":"aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83"} Oct 03 13:53:23 crc kubenswrapper[4959]: I1003 13:53:23.804564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerStarted","Data":"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a"} Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.819183 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerStarted","Data":"342111745eadc22d2c2eafdf7a166b38b56b789123de46efd4bcb84e87205c69"} Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.824184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" event={"ID":"b13a531c-8bcc-46b4-b034-525a5774081d","Type":"ContainerStarted","Data":"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942"} Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.824329 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.826552 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerStarted","Data":"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a"} Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.826680 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api-log" containerID="cri-o://4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" gracePeriod=30 Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.826743 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api" containerID="cri-o://3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" gracePeriod=30 Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.826802 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.848423 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" podStartSLOduration=4.848402451 podStartE2EDuration="4.848402451s" podCreationTimestamp="2025-10-03 13:53:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:24.843901912 +0000 UTC m=+1374.047245329" watchObservedRunningTime="2025-10-03 13:53:24.848402451 +0000 UTC m=+1374.051745868" Oct 03 13:53:24 crc kubenswrapper[4959]: I1003 13:53:24.871174 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.871156634 podStartE2EDuration="4.871156634s" podCreationTimestamp="2025-10-03 13:53:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:24.862438761 +0000 UTC m=+1374.065782178" watchObservedRunningTime="2025-10-03 13:53:24.871156634 +0000 UTC m=+1374.074500041" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.072251 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-t4tw9"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.073788 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.094897 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t4tw9"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.155212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl9bp\" (UniqueName: \"kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp\") pod \"nova-api-db-create-t4tw9\" (UID: \"dbc3e407-060d-4f10-8993-71f733e510ed\") " pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.172508 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-p2q4s"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.174027 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.217900 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p2q4s"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.258367 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dh4v\" (UniqueName: \"kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v\") pod \"nova-cell0-db-create-p2q4s\" (UID: \"7830ad44-ea59-47a8-973b-4438930500dc\") " pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.258472 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl9bp\" (UniqueName: \"kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp\") pod \"nova-api-db-create-t4tw9\" (UID: \"dbc3e407-060d-4f10-8993-71f733e510ed\") " pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.260762 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.261171 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-log" containerID="cri-o://8c164c5247a2631e33c76e1ec7e46302a319f9bf2b75393b0785b327b5b1ed77" gracePeriod=30 Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.262355 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-httpd" containerID="cri-o://87681968bf3b43d810f5c797f220cf83f293f5af560789c5caf22b22226ad5c3" gracePeriod=30 Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.296181 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl9bp\" (UniqueName: \"kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp\") pod \"nova-api-db-create-t4tw9\" (UID: \"dbc3e407-060d-4f10-8993-71f733e510ed\") " pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.360418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dh4v\" (UniqueName: \"kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v\") pod \"nova-cell0-db-create-p2q4s\" (UID: \"7830ad44-ea59-47a8-973b-4438930500dc\") " pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.371153 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b4php"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.392803 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.394506 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.395940 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b4php"] Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.400148 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dh4v\" (UniqueName: \"kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v\") pod \"nova-cell0-db-create-p2q4s\" (UID: \"7830ad44-ea59-47a8-973b-4438930500dc\") " pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.468116 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcl72\" (UniqueName: \"kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72\") pod \"nova-cell1-db-create-b4php\" (UID: \"58ba97c9-9be2-41bd-a165-e3baecf9c97a\") " pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.511831 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.581057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcl72\" (UniqueName: \"kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72\") pod \"nova-cell1-db-create-b4php\" (UID: \"58ba97c9-9be2-41bd-a165-e3baecf9c97a\") " pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.603765 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcl72\" (UniqueName: \"kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72\") pod \"nova-cell1-db-create-b4php\" (UID: \"58ba97c9-9be2-41bd-a165-e3baecf9c97a\") " pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.754441 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.811204 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.862692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerStarted","Data":"8d941f6234f05f739d47faef13a939ca11b37cebbc5d138c109d44721c4cf7ff"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.868004 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerStarted","Data":"1687cfe1fb627021244443cd8feaa5d63671fc0f72f8a818a5dbd7ec7dc6d2a7"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895065 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895246 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895286 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895347 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895411 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895443 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55dq\" (UniqueName: \"kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.895613 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id\") pod \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\" (UID: \"687cdb29-cb8a-47e6-b5a7-a4237b15d7de\") " Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.897401 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs" (OuterVolumeSpecName: "logs") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.897724 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.897803 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.902971 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts" (OuterVolumeSpecName: "scripts") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.903020 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.905395 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq" (OuterVolumeSpecName: "kube-api-access-x55dq") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "kube-api-access-x55dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.930729 4959 generic.go:334] "Generic (PLEG): container finished" podID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerID="8c164c5247a2631e33c76e1ec7e46302a319f9bf2b75393b0785b327b5b1ed77" exitCode=143 Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.930868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerDied","Data":"8c164c5247a2631e33c76e1ec7e46302a319f9bf2b75393b0785b327b5b1ed77"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.933282 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.954710 4959 generic.go:334] "Generic (PLEG): container finished" podID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerID="3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" exitCode=0 Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.954751 4959 generic.go:334] "Generic (PLEG): container finished" podID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerID="4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" exitCode=143 Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.955924 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.956495 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerDied","Data":"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.956533 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerDied","Data":"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.956548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"687cdb29-cb8a-47e6-b5a7-a4237b15d7de","Type":"ContainerDied","Data":"2d2285819d300c3728cf7e45ef5f5afa78cd7d40ed582f6081484149703a2428"} Oct 03 13:53:25 crc kubenswrapper[4959]: I1003 13:53:25.956566 4959 scope.go:117] "RemoveContainer" containerID="3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.003356 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.003383 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.003394 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55dq\" (UniqueName: \"kubernetes.io/projected/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-kube-api-access-x55dq\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.003403 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.003413 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.020314 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data" (OuterVolumeSpecName: "config-data") pod "687cdb29-cb8a-47e6-b5a7-a4237b15d7de" (UID: "687cdb29-cb8a-47e6-b5a7-a4237b15d7de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.111396 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t4tw9"] Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.113469 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687cdb29-cb8a-47e6-b5a7-a4237b15d7de-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.122502 4959 scope.go:117] "RemoveContainer" containerID="4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.253786 4959 scope.go:117] "RemoveContainer" containerID="3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" Oct 03 13:53:26 crc kubenswrapper[4959]: E1003 13:53:26.254340 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a\": container with ID starting with 3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a not found: ID does not exist" containerID="3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254381 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a"} err="failed to get container status \"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a\": rpc error: code = NotFound desc = could not find container \"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a\": container with ID starting with 3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a not found: ID does not exist" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254407 4959 scope.go:117] "RemoveContainer" containerID="4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" Oct 03 13:53:26 crc kubenswrapper[4959]: E1003 13:53:26.254660 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a\": container with ID starting with 4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a not found: ID does not exist" containerID="4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254693 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a"} err="failed to get container status \"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a\": rpc error: code = NotFound desc = could not find container \"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a\": container with ID starting with 4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a not found: ID does not exist" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254714 4959 scope.go:117] "RemoveContainer" containerID="3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254933 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a"} err="failed to get container status \"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a\": rpc error: code = NotFound desc = could not find container \"3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a\": container with ID starting with 3ec6db8e84ba7fc68c87fce8911ac473c03c59c8c4c56d31e213d74d649ad92a not found: ID does not exist" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.254947 4959 scope.go:117] "RemoveContainer" containerID="4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.255089 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a"} err="failed to get container status \"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a\": rpc error: code = NotFound desc = could not find container \"4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a\": container with ID starting with 4503ef0df3c153195368c9e2b65ff4383767e5e41dd372e2a54f4b0f2493b89a not found: ID does not exist" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.269082 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p2q4s"] Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.314258 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.341265 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.358433 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:26 crc kubenswrapper[4959]: E1003 13:53:26.358871 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.358893 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api" Oct 03 13:53:26 crc kubenswrapper[4959]: E1003 13:53:26.358909 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api-log" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.358918 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api-log" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.359152 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.359212 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" containerName="cinder-api-log" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.360316 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.365911 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.366176 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.370003 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.386900 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.478764 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b4php"] Oct 03 13:53:26 crc kubenswrapper[4959]: W1003 13:53:26.492262 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58ba97c9_9be2_41bd_a165_e3baecf9c97a.slice/crio-199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de WatchSource:0}: Error finding container 199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de: Status 404 returned error can't find the container with id 199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526220 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526255 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6x7d\" (UniqueName: \"kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526514 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526532 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526552 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526582 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526598 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.526989 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.628770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629027 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629148 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629310 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629440 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.628883 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.629995 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.630142 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.630324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6x7d\" (UniqueName: \"kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.630252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.634637 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.639732 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.640792 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.643663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.643914 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.645233 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.652207 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6x7d\" (UniqueName: \"kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d\") pod \"cinder-api-0\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " pod="openstack/cinder-api-0" Oct 03 13:53:26 crc kubenswrapper[4959]: I1003 13:53:26.778509 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.034150 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p2q4s" event={"ID":"7830ad44-ea59-47a8-973b-4438930500dc","Type":"ContainerStarted","Data":"8ad815a404d34e7daeb62fd5d97fc17482809b2f82c91a7f5724b4b1e3321266"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.034556 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p2q4s" event={"ID":"7830ad44-ea59-47a8-973b-4438930500dc","Type":"ContainerStarted","Data":"7e736f2885b1893d843fdffbe979eb10d81a6ca08a6d87db7c2e5915f800fa77"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.043740 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerStarted","Data":"0c3551e59af6fc5ab2f8821441f09b2e4f1a3e638bc16e383d5894343df5e70e"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.048858 4959 generic.go:334] "Generic (PLEG): container finished" podID="dbc3e407-060d-4f10-8993-71f733e510ed" containerID="a7bbb7b660035d1128449ae672bafdccb24dae017043de5c764061910cfda76f" exitCode=0 Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.048933 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t4tw9" event={"ID":"dbc3e407-060d-4f10-8993-71f733e510ed","Type":"ContainerDied","Data":"a7bbb7b660035d1128449ae672bafdccb24dae017043de5c764061910cfda76f"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.048962 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t4tw9" event={"ID":"dbc3e407-060d-4f10-8993-71f733e510ed","Type":"ContainerStarted","Data":"ad62afaf86b609f3551ba0ba4e306a7b32b9c0eda29f8ac41ae25b41c2b1ad06"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.064821 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b4php" event={"ID":"58ba97c9-9be2-41bd-a165-e3baecf9c97a","Type":"ContainerStarted","Data":"5c237f828a85840621666072b2ca9c1f967a6c52a22efbea5023d5b7c68c4fbc"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.064869 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b4php" event={"ID":"58ba97c9-9be2-41bd-a165-e3baecf9c97a","Type":"ContainerStarted","Data":"199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de"} Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.077157 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-p2q4s" podStartSLOduration=2.077135386 podStartE2EDuration="2.077135386s" podCreationTimestamp="2025-10-03 13:53:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:27.055609954 +0000 UTC m=+1376.258953381" watchObservedRunningTime="2025-10-03 13:53:27.077135386 +0000 UTC m=+1376.280478803" Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.098975 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.273941289 podStartE2EDuration="8.098954696s" podCreationTimestamp="2025-10-03 13:53:19 +0000 UTC" firstStartedPulling="2025-10-03 13:53:22.457595164 +0000 UTC m=+1371.660938581" lastFinishedPulling="2025-10-03 13:53:24.282608571 +0000 UTC m=+1373.485951988" observedRunningTime="2025-10-03 13:53:27.096831534 +0000 UTC m=+1376.300174971" watchObservedRunningTime="2025-10-03 13:53:27.098954696 +0000 UTC m=+1376.302298103" Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.116411 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-b4php" podStartSLOduration=2.116393298 podStartE2EDuration="2.116393298s" podCreationTimestamp="2025-10-03 13:53:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:27.111889209 +0000 UTC m=+1376.315232636" watchObservedRunningTime="2025-10-03 13:53:27.116393298 +0000 UTC m=+1376.319736715" Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.341170 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:53:27 crc kubenswrapper[4959]: W1003 13:53:27.345972 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71c2bd3a_c4a2_4ad1_bfce_85f0194dfb1a.slice/crio-1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3 WatchSource:0}: Error finding container 1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3: Status 404 returned error can't find the container with id 1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3 Oct 03 13:53:27 crc kubenswrapper[4959]: I1003 13:53:27.704508 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="687cdb29-cb8a-47e6-b5a7-a4237b15d7de" path="/var/lib/kubelet/pods/687cdb29-cb8a-47e6-b5a7-a4237b15d7de/volumes" Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.080649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerStarted","Data":"b310fb69f996618095e2ad2c4d0857f61089302454ce26983857cb17725a0561"} Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.092109 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerStarted","Data":"cc4b4c055039a16bb8698b1f062f4ab7a7185cd1b62caa23f240835b1d1163dd"} Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.092257 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerStarted","Data":"1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3"} Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.093868 4959 generic.go:334] "Generic (PLEG): container finished" podID="58ba97c9-9be2-41bd-a165-e3baecf9c97a" containerID="5c237f828a85840621666072b2ca9c1f967a6c52a22efbea5023d5b7c68c4fbc" exitCode=0 Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.093970 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b4php" event={"ID":"58ba97c9-9be2-41bd-a165-e3baecf9c97a","Type":"ContainerDied","Data":"5c237f828a85840621666072b2ca9c1f967a6c52a22efbea5023d5b7c68c4fbc"} Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.095427 4959 generic.go:334] "Generic (PLEG): container finished" podID="7830ad44-ea59-47a8-973b-4438930500dc" containerID="8ad815a404d34e7daeb62fd5d97fc17482809b2f82c91a7f5724b4b1e3321266" exitCode=0 Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.095571 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p2q4s" event={"ID":"7830ad44-ea59-47a8-973b-4438930500dc","Type":"ContainerDied","Data":"8ad815a404d34e7daeb62fd5d97fc17482809b2f82c91a7f5724b4b1e3321266"} Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.351325 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.351614 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-log" containerID="cri-o://6f65f901047f072024008154bcb5706dd05c5c1c069a77f6cce157da11e35684" gracePeriod=30 Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.351766 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-httpd" containerID="cri-o://99ae8be8736d1264a4e5a30b3acec9a9ccd988ea8c5766acc11621fb435ee82c" gracePeriod=30 Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.588691 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.672923 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl9bp\" (UniqueName: \"kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp\") pod \"dbc3e407-060d-4f10-8993-71f733e510ed\" (UID: \"dbc3e407-060d-4f10-8993-71f733e510ed\") " Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.681314 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp" (OuterVolumeSpecName: "kube-api-access-cl9bp") pod "dbc3e407-060d-4f10-8993-71f733e510ed" (UID: "dbc3e407-060d-4f10-8993-71f733e510ed"). InnerVolumeSpecName "kube-api-access-cl9bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:28 crc kubenswrapper[4959]: I1003 13:53:28.777630 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl9bp\" (UniqueName: \"kubernetes.io/projected/dbc3e407-060d-4f10-8993-71f733e510ed-kube-api-access-cl9bp\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.106873 4959 generic.go:334] "Generic (PLEG): container finished" podID="9cea2d9b-538b-4333-a81f-c59792924b06" containerID="6f65f901047f072024008154bcb5706dd05c5c1c069a77f6cce157da11e35684" exitCode=143 Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.107172 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerDied","Data":"6f65f901047f072024008154bcb5706dd05c5c1c069a77f6cce157da11e35684"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.111023 4959 generic.go:334] "Generic (PLEG): container finished" podID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerID="87681968bf3b43d810f5c797f220cf83f293f5af560789c5caf22b22226ad5c3" exitCode=0 Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.111076 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerDied","Data":"87681968bf3b43d810f5c797f220cf83f293f5af560789c5caf22b22226ad5c3"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.111101 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7f301c6f-50dd-410f-b6cf-9d62d154289e","Type":"ContainerDied","Data":"895ba025e112e38efeebd5087798538e71bd491fd6f4aa9c4ad1eec72c06f4a8"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.111111 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895ba025e112e38efeebd5087798538e71bd491fd6f4aa9c4ad1eec72c06f4a8" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.113626 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerStarted","Data":"a7c1d4fb80394150b035616306b0249611184085fee535c8c9ed09e63d2e3572"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.115511 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.119530 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerStarted","Data":"ee16252e1bbddaf8d1e22642807ff2d9e9259f29027e5a4fff2c75e96b01eb37"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.120533 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.122760 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t4tw9" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.122863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t4tw9" event={"ID":"dbc3e407-060d-4f10-8993-71f733e510ed","Type":"ContainerDied","Data":"ad62afaf86b609f3551ba0ba4e306a7b32b9c0eda29f8ac41ae25b41c2b1ad06"} Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.124021 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad62afaf86b609f3551ba0ba4e306a7b32b9c0eda29f8ac41ae25b41c2b1ad06" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.149921 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.1498993459999998 podStartE2EDuration="3.149899346s" podCreationTimestamp="2025-10-03 13:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:29.135027445 +0000 UTC m=+1378.338370872" watchObservedRunningTime="2025-10-03 13:53:29.149899346 +0000 UTC m=+1378.353242763" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.206619 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.233811 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.246270873 podStartE2EDuration="9.233795172s" podCreationTimestamp="2025-10-03 13:53:20 +0000 UTC" firstStartedPulling="2025-10-03 13:53:22.626623736 +0000 UTC m=+1371.829967163" lastFinishedPulling="2025-10-03 13:53:28.614148045 +0000 UTC m=+1377.817491462" observedRunningTime="2025-10-03 13:53:29.170642268 +0000 UTC m=+1378.373985685" watchObservedRunningTime="2025-10-03 13:53:29.233795172 +0000 UTC m=+1378.437138589" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.286946 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.286995 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287014 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287048 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6cvb\" (UniqueName: \"kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287105 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287222 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287295 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.287348 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle\") pod \"7f301c6f-50dd-410f-b6cf-9d62d154289e\" (UID: \"7f301c6f-50dd-410f-b6cf-9d62d154289e\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.288287 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.297026 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs" (OuterVolumeSpecName: "logs") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.298240 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.298371 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts" (OuterVolumeSpecName: "scripts") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.313699 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb" (OuterVolumeSpecName: "kube-api-access-g6cvb") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "kube-api-access-g6cvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.323380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.352632 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data" (OuterVolumeSpecName: "config-data") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.372378 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7f301c6f-50dd-410f-b6cf-9d62d154289e" (UID: "7f301c6f-50dd-410f-b6cf-9d62d154289e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390102 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390142 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390180 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390209 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390222 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390232 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f301c6f-50dd-410f-b6cf-9d62d154289e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390242 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f301c6f-50dd-410f-b6cf-9d62d154289e-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.390253 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6cvb\" (UniqueName: \"kubernetes.io/projected/7f301c6f-50dd-410f-b6cf-9d62d154289e-kube-api-access-g6cvb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.420586 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.493230 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.603734 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.695044 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.696049 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dh4v\" (UniqueName: \"kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v\") pod \"7830ad44-ea59-47a8-973b-4438930500dc\" (UID: \"7830ad44-ea59-47a8-973b-4438930500dc\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.707577 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v" (OuterVolumeSpecName: "kube-api-access-6dh4v") pod "7830ad44-ea59-47a8-973b-4438930500dc" (UID: "7830ad44-ea59-47a8-973b-4438930500dc"). InnerVolumeSpecName "kube-api-access-6dh4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.797973 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcl72\" (UniqueName: \"kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72\") pod \"58ba97c9-9be2-41bd-a165-e3baecf9c97a\" (UID: \"58ba97c9-9be2-41bd-a165-e3baecf9c97a\") " Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.798668 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dh4v\" (UniqueName: \"kubernetes.io/projected/7830ad44-ea59-47a8-973b-4438930500dc-kube-api-access-6dh4v\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.805427 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72" (OuterVolumeSpecName: "kube-api-access-zcl72") pod "58ba97c9-9be2-41bd-a165-e3baecf9c97a" (UID: "58ba97c9-9be2-41bd-a165-e3baecf9c97a"). InnerVolumeSpecName "kube-api-access-zcl72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:29 crc kubenswrapper[4959]: I1003 13:53:29.900424 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcl72\" (UniqueName: \"kubernetes.io/projected/58ba97c9-9be2-41bd-a165-e3baecf9c97a-kube-api-access-zcl72\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.133022 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b4php" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.133012 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b4php" event={"ID":"58ba97c9-9be2-41bd-a165-e3baecf9c97a","Type":"ContainerDied","Data":"199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de"} Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.134221 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="199be659c839798c0496e53e1b176b9a367ebf81dd4711d86881bbb79dc3f5de" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.135868 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p2q4s" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.135896 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.145682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p2q4s" event={"ID":"7830ad44-ea59-47a8-973b-4438930500dc","Type":"ContainerDied","Data":"7e736f2885b1893d843fdffbe979eb10d81a6ca08a6d87db7c2e5915f800fa77"} Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.145791 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e736f2885b1893d843fdffbe979eb10d81a6ca08a6d87db7c2e5915f800fa77" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.181562 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.191022 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.202981 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:30 crc kubenswrapper[4959]: E1003 13:53:30.203376 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ba97c9-9be2-41bd-a165-e3baecf9c97a" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203392 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ba97c9-9be2-41bd-a165-e3baecf9c97a" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: E1003 13:53:30.203401 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-httpd" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203409 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-httpd" Oct 03 13:53:30 crc kubenswrapper[4959]: E1003 13:53:30.203419 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-log" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203425 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-log" Oct 03 13:53:30 crc kubenswrapper[4959]: E1003 13:53:30.203432 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc3e407-060d-4f10-8993-71f733e510ed" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203438 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc3e407-060d-4f10-8993-71f733e510ed" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: E1003 13:53:30.203451 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7830ad44-ea59-47a8-973b-4438930500dc" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203457 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7830ad44-ea59-47a8-973b-4438930500dc" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203637 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ba97c9-9be2-41bd-a165-e3baecf9c97a" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203668 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc3e407-060d-4f10-8993-71f733e510ed" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203682 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-httpd" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203696 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7830ad44-ea59-47a8-973b-4438930500dc" containerName="mariadb-database-create" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.203705 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" containerName="glance-log" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.204586 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.211708 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.219813 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.228803 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.277482 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307014 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307529 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d5fl\" (UniqueName: \"kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307763 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307846 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.307963 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.308038 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.308114 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.403572 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409654 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409711 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409799 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409829 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.409908 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d5fl\" (UniqueName: \"kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.410017 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.410042 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.410665 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.411078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.411147 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.415457 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.416915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.417385 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.418014 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.429542 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d5fl\" (UniqueName: \"kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.486923 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.518762 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.523609 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.536626 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.536889 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="dnsmasq-dns" containerID="cri-o://9420f8084d458dabaa75af29ba5a0a48d6ea4846db169c617c705cda323e00ef" gracePeriod=10 Oct 03 13:53:30 crc kubenswrapper[4959]: I1003 13:53:30.701228 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.153352 4959 generic.go:334] "Generic (PLEG): container finished" podID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerID="9420f8084d458dabaa75af29ba5a0a48d6ea4846db169c617c705cda323e00ef" exitCode=0 Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.153803 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" event={"ID":"ccca3d9a-e8e7-4699-8713-21d6ce4445c1","Type":"ContainerDied","Data":"9420f8084d458dabaa75af29ba5a0a48d6ea4846db169c617c705cda323e00ef"} Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.272388 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.278284 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.307029 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.332443 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.333254 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.333339 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.333611 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.333650 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkpkf\" (UniqueName: \"kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.333716 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config\") pod \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\" (UID: \"ccca3d9a-e8e7-4699-8713-21d6ce4445c1\") " Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.341847 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf" (OuterVolumeSpecName: "kube-api-access-jkpkf") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "kube-api-access-jkpkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.391285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.435482 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.435806 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkpkf\" (UniqueName: \"kubernetes.io/projected/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-kube-api-access-jkpkf\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.436739 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.437126 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.451082 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config" (OuterVolumeSpecName: "config") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.457041 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccca3d9a-e8e7-4699-8713-21d6ce4445c1" (UID: "ccca3d9a-e8e7-4699-8713-21d6ce4445c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.537411 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.537439 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.537449 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.537481 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccca3d9a-e8e7-4699-8713-21d6ce4445c1-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.591511 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:43340->10.217.0.153:9292: read: connection reset by peer" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.591580 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:43350->10.217.0.153:9292: read: connection reset by peer" Oct 03 13:53:31 crc kubenswrapper[4959]: I1003 13:53:31.715495 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f301c6f-50dd-410f-b6cf-9d62d154289e" path="/var/lib/kubelet/pods/7f301c6f-50dd-410f-b6cf-9d62d154289e/volumes" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.179839 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" event={"ID":"ccca3d9a-e8e7-4699-8713-21d6ce4445c1","Type":"ContainerDied","Data":"389a8010bc94b8a851e03b2781d4f8c794eec965742f3e8c5e2acada0c988934"} Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.180118 4959 scope.go:117] "RemoveContainer" containerID="9420f8084d458dabaa75af29ba5a0a48d6ea4846db169c617c705cda323e00ef" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.180312 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-xm82z" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.185177 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerStarted","Data":"f2500fe9d0158d0ab935afc5fe14db033c74992aa7ae26ecba958ae3596c57f9"} Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.185245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerStarted","Data":"924b322a592049951ad85ab4516707758a112c97bd941a0fd5404e90ff4ca8c5"} Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.197163 4959 generic.go:334] "Generic (PLEG): container finished" podID="9cea2d9b-538b-4333-a81f-c59792924b06" containerID="99ae8be8736d1264a4e5a30b3acec9a9ccd988ea8c5766acc11621fb435ee82c" exitCode=0 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.197560 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-central-agent" containerID="cri-o://342111745eadc22d2c2eafdf7a166b38b56b789123de46efd4bcb84e87205c69" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.197880 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerDied","Data":"99ae8be8736d1264a4e5a30b3acec9a9ccd988ea8c5766acc11621fb435ee82c"} Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.197929 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="proxy-httpd" containerID="cri-o://ee16252e1bbddaf8d1e22642807ff2d9e9259f29027e5a4fff2c75e96b01eb37" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.197985 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="sg-core" containerID="cri-o://b310fb69f996618095e2ad2c4d0857f61089302454ce26983857cb17725a0561" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.198042 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-notification-agent" containerID="cri-o://1687cfe1fb627021244443cd8feaa5d63671fc0f72f8a818a5dbd7ec7dc6d2a7" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.198388 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="probe" containerID="cri-o://0c3551e59af6fc5ab2f8821441f09b2e4f1a3e638bc16e383d5894343df5e70e" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.198473 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="cinder-scheduler" containerID="cri-o://8d941f6234f05f739d47faef13a939ca11b37cebbc5d138c109d44721c4cf7ff" gracePeriod=30 Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.262044 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.278631 4959 scope.go:117] "RemoveContainer" containerID="812e6e0224c1b813604ba1dc36cb6511db46d98c9630a312f4db5a4e6c19e769" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.289901 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.315520 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-xm82z"] Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.349743 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.349918 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.349942 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.350010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.350041 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mtnm\" (UniqueName: \"kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.350108 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.350201 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.350234 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"9cea2d9b-538b-4333-a81f-c59792924b06\" (UID: \"9cea2d9b-538b-4333-a81f-c59792924b06\") " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.353754 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.354111 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs" (OuterVolumeSpecName: "logs") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.368400 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts" (OuterVolumeSpecName: "scripts") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.370559 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.372581 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm" (OuterVolumeSpecName: "kube-api-access-4mtnm") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "kube-api-access-4mtnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.392961 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.430763 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452436 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452490 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452502 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452511 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452520 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452554 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mtnm\" (UniqueName: \"kubernetes.io/projected/9cea2d9b-538b-4333-a81f-c59792924b06-kube-api-access-4mtnm\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.452564 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cea2d9b-538b-4333-a81f-c59792924b06-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.472686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data" (OuterVolumeSpecName: "config-data") pod "9cea2d9b-538b-4333-a81f-c59792924b06" (UID: "9cea2d9b-538b-4333-a81f-c59792924b06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.492952 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.554149 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.554181 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cea2d9b-538b-4333-a81f-c59792924b06-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:32 crc kubenswrapper[4959]: I1003 13:53:32.654768 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.207864 4959 generic.go:334] "Generic (PLEG): container finished" podID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerID="0c3551e59af6fc5ab2f8821441f09b2e4f1a3e638bc16e383d5894343df5e70e" exitCode=0 Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.208219 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerDied","Data":"0c3551e59af6fc5ab2f8821441f09b2e4f1a3e638bc16e383d5894343df5e70e"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211706 4959 generic.go:334] "Generic (PLEG): container finished" podID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerID="ee16252e1bbddaf8d1e22642807ff2d9e9259f29027e5a4fff2c75e96b01eb37" exitCode=0 Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211746 4959 generic.go:334] "Generic (PLEG): container finished" podID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerID="b310fb69f996618095e2ad2c4d0857f61089302454ce26983857cb17725a0561" exitCode=2 Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211757 4959 generic.go:334] "Generic (PLEG): container finished" podID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerID="1687cfe1fb627021244443cd8feaa5d63671fc0f72f8a818a5dbd7ec7dc6d2a7" exitCode=0 Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211766 4959 generic.go:334] "Generic (PLEG): container finished" podID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerID="342111745eadc22d2c2eafdf7a166b38b56b789123de46efd4bcb84e87205c69" exitCode=0 Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211773 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerDied","Data":"ee16252e1bbddaf8d1e22642807ff2d9e9259f29027e5a4fff2c75e96b01eb37"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211803 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerDied","Data":"b310fb69f996618095e2ad2c4d0857f61089302454ce26983857cb17725a0561"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211815 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerDied","Data":"1687cfe1fb627021244443cd8feaa5d63671fc0f72f8a818a5dbd7ec7dc6d2a7"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211823 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerDied","Data":"342111745eadc22d2c2eafdf7a166b38b56b789123de46efd4bcb84e87205c69"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211832 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5","Type":"ContainerDied","Data":"174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.211842 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="174cf52bb0a9e10b59d217527ddf3c83a19205ab26a0ace732f6a7f20463d77a" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.214180 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerStarted","Data":"fc9d972ac9b2f06f0e0356aa259d27ac5d465cfc53bdfcc611986b0718292921"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.216063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9cea2d9b-538b-4333-a81f-c59792924b06","Type":"ContainerDied","Data":"64af77666280070d4b3c2e94d1891640387ed89f27033f521e471b6905e84a84"} Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.216144 4959 scope.go:117] "RemoveContainer" containerID="99ae8be8736d1264a4e5a30b3acec9a9ccd988ea8c5766acc11621fb435ee82c" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.216155 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.294981 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.313739 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.316411 4959 scope.go:117] "RemoveContainer" containerID="6f65f901047f072024008154bcb5706dd05c5c1c069a77f6cce157da11e35684" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.337267 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.354840 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355339 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="sg-core" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355358 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="sg-core" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355375 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-log" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355383 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-log" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355408 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-notification-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355416 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-notification-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355431 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="init" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355438 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="init" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355456 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="proxy-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355464 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="proxy-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355484 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355492 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355505 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-central-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355512 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-central-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: E1003 13:53:33.355531 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="dnsmasq-dns" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355537 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="dnsmasq-dns" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355754 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="sg-core" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355772 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-central-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355784 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="proxy-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355801 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-httpd" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355814 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" containerName="glance-log" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355826 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" containerName="ceilometer-notification-agent" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.355843 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" containerName="dnsmasq-dns" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.357182 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.360130 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.360359 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.371802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.371918 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.371982 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.372080 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h45qr\" (UniqueName: \"kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.372107 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.372137 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.372173 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd\") pod \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\" (UID: \"c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5\") " Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.372630 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.375338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.386489 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.391441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr" (OuterVolumeSpecName: "kube-api-access-h45qr") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "kube-api-access-h45qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.416330 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts" (OuterVolumeSpecName: "scripts") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.433607 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.474518 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.474828 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.474974 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.475243 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.475480 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.475743 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.475897 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476105 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw5gn\" (UniqueName: \"kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476313 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476499 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476624 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h45qr\" (UniqueName: \"kubernetes.io/projected/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-kube-api-access-h45qr\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476739 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.476848 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.509648 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.521160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data" (OuterVolumeSpecName: "config-data") pod "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" (UID: "c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.578601 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.578868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.579383 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.579334 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580118 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580252 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580404 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580533 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580659 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw5gn\" (UniqueName: \"kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580657 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580861 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580833 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.580898 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.583545 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.584034 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.584267 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.595687 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.598644 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw5gn\" (UniqueName: \"kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.609365 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.696296 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cea2d9b-538b-4333-a81f-c59792924b06" path="/var/lib/kubelet/pods/9cea2d9b-538b-4333-a81f-c59792924b06/volumes" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.696854 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccca3d9a-e8e7-4699-8713-21d6ce4445c1" path="/var/lib/kubelet/pods/ccca3d9a-e8e7-4699-8713-21d6ce4445c1/volumes" Oct 03 13:53:33 crc kubenswrapper[4959]: I1003 13:53:33.799427 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.233981 4959 generic.go:334] "Generic (PLEG): container finished" podID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerID="8d941f6234f05f739d47faef13a939ca11b37cebbc5d138c109d44721c4cf7ff" exitCode=0 Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.234127 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerDied","Data":"8d941f6234f05f739d47faef13a939ca11b37cebbc5d138c109d44721c4cf7ff"} Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.234335 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.267296 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.267279469 podStartE2EDuration="4.267279469s" podCreationTimestamp="2025-10-03 13:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:34.261825107 +0000 UTC m=+1383.465168524" watchObservedRunningTime="2025-10-03 13:53:34.267279469 +0000 UTC m=+1383.470622886" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.321030 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.344360 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.365273 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.384449 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.390904 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396017 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396090 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396115 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396122 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396142 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmjzv\" (UniqueName: \"kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396227 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.396253 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.397816 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.422071 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498427 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498465 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmjzv\" (UniqueName: \"kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.498482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.499981 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.502414 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.506036 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.506860 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.507428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.508438 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.522097 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmjzv\" (UniqueName: \"kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv\") pod \"ceilometer-0\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.574666 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.585446 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600112 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600221 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600253 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvjbs\" (UniqueName: \"kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600255 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600327 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.600408 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data\") pod \"a3c31694-94d2-4362-ac72-c336f3b334b8\" (UID: \"a3c31694-94d2-4362-ac72-c336f3b334b8\") " Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.601001 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c31694-94d2-4362-ac72-c336f3b334b8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.603381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.605377 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts" (OuterVolumeSpecName: "scripts") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.611413 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs" (OuterVolumeSpecName: "kube-api-access-kvjbs") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "kube-api-access-kvjbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.673929 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.702599 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.702633 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.702645 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvjbs\" (UniqueName: \"kubernetes.io/projected/a3c31694-94d2-4362-ac72-c336f3b334b8-kube-api-access-kvjbs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.702659 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.724882 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data" (OuterVolumeSpecName: "config-data") pod "a3c31694-94d2-4362-ac72-c336f3b334b8" (UID: "a3c31694-94d2-4362-ac72-c336f3b334b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.805124 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c31694-94d2-4362-ac72-c336f3b334b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.920649 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.982244 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.982464 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-596475bc7b-44mtq" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-api" containerID="cri-o://102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc" gracePeriod=30 Oct 03 13:53:34 crc kubenswrapper[4959]: I1003 13:53:34.982847 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-596475bc7b-44mtq" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-httpd" containerID="cri-o://9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d" gracePeriod=30 Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.095330 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.125680 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3cd4-account-create-6tgxr"] Oct 03 13:53:35 crc kubenswrapper[4959]: E1003 13:53:35.126441 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="probe" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.126459 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="probe" Oct 03 13:53:35 crc kubenswrapper[4959]: E1003 13:53:35.126491 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="cinder-scheduler" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.126499 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="cinder-scheduler" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.126676 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="cinder-scheduler" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.126687 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" containerName="probe" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.127245 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.129180 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.152241 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3cd4-account-create-6tgxr"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.247395 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerStarted","Data":"3dd937b5c4e769e9e6ce845a3dcff4ee73580495bb9019b6917cda690b07df00"} Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.250519 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerStarted","Data":"a7b0ceeff61941890b21a6f7d2ca35938e4cdfd15247925f0209a5119fb4f1a1"} Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.250547 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerStarted","Data":"10e3f5d56da42bee734ca1370b0c5cbc404d0a15225c6755fdb4a9e65b84865a"} Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.252164 4959 generic.go:334] "Generic (PLEG): container finished" podID="47d02229-11f4-4ad4-ad25-698800155fdb" containerID="9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d" exitCode=0 Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.252239 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerDied","Data":"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d"} Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.254646 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c31694-94d2-4362-ac72-c336f3b334b8","Type":"ContainerDied","Data":"8e7a71c916d834fbba63543e377476859b078486bef96e681fea5f4f3333bd2a"} Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.254683 4959 scope.go:117] "RemoveContainer" containerID="0c3551e59af6fc5ab2f8821441f09b2e4f1a3e638bc16e383d5894343df5e70e" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.254718 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.313450 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.317967 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t8tc\" (UniqueName: \"kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc\") pod \"nova-api-3cd4-account-create-6tgxr\" (UID: \"cdc575e3-fc82-48e4-ae54-27f374e37063\") " pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.318235 4959 scope.go:117] "RemoveContainer" containerID="8d941f6234f05f739d47faef13a939ca11b37cebbc5d138c109d44721c4cf7ff" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.342999 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.357146 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.359111 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.361677 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.367228 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.419980 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420080 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420120 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420159 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f6gf\" (UniqueName: \"kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420181 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420307 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.420385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t8tc\" (UniqueName: \"kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc\") pod \"nova-api-3cd4-account-create-6tgxr\" (UID: \"cdc575e3-fc82-48e4-ae54-27f374e37063\") " pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.456775 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t8tc\" (UniqueName: \"kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc\") pod \"nova-api-3cd4-account-create-6tgxr\" (UID: \"cdc575e3-fc82-48e4-ae54-27f374e37063\") " pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.519798 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521847 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521899 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521945 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f6gf\" (UniqueName: \"kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521965 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.521991 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.522152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.525442 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.528705 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.529102 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.529351 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.538778 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f6gf\" (UniqueName: \"kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf\") pod \"cinder-scheduler-0\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.681049 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.703332 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3c31694-94d2-4362-ac72-c336f3b334b8" path="/var/lib/kubelet/pods/a3c31694-94d2-4362-ac72-c336f3b334b8/volumes" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.704347 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5" path="/var/lib/kubelet/pods/c1cfe967-4129-4b6b-a6e9-1ec8c96a00d5/volumes" Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.875822 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:35 crc kubenswrapper[4959]: I1003 13:53:35.959512 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3cd4-account-create-6tgxr"] Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.180494 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:53:36 crc kubenswrapper[4959]: W1003 13:53:36.190538 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2f9c520_72f3_482f_be82_8a625a24f3e1.slice/crio-e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2 WatchSource:0}: Error finding container e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2: Status 404 returned error can't find the container with id e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2 Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.264496 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerStarted","Data":"e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2"} Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.269211 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerStarted","Data":"cef4485dd143a31defd4cf196d86695d90b172ab45acac529670496587ccfe92"} Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.273563 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerStarted","Data":"9bc9b789b1e410f14438dd275ab4df1b4886b1a5b73a1ba9e3ec567e82067001"} Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.277504 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cd4-account-create-6tgxr" event={"ID":"cdc575e3-fc82-48e4-ae54-27f374e37063","Type":"ContainerStarted","Data":"301f13bc924830f186afba1bebfa0f509472a654adc102a985467ad83327e434"} Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.277550 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cd4-account-create-6tgxr" event={"ID":"cdc575e3-fc82-48e4-ae54-27f374e37063","Type":"ContainerStarted","Data":"f50c54a3c106c5171991105a105452894a8795d9cf98f093622d0a4128570dd2"} Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.298432 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.298412748 podStartE2EDuration="3.298412748s" podCreationTimestamp="2025-10-03 13:53:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:36.296526173 +0000 UTC m=+1385.499869610" watchObservedRunningTime="2025-10-03 13:53:36.298412748 +0000 UTC m=+1385.501756165" Oct 03 13:53:36 crc kubenswrapper[4959]: I1003 13:53:36.338575 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-3cd4-account-create-6tgxr" podStartSLOduration=1.338552032 podStartE2EDuration="1.338552032s" podCreationTimestamp="2025-10-03 13:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:36.316655701 +0000 UTC m=+1385.519999128" watchObservedRunningTime="2025-10-03 13:53:36.338552032 +0000 UTC m=+1385.541895449" Oct 03 13:53:37 crc kubenswrapper[4959]: I1003 13:53:37.300272 4959 generic.go:334] "Generic (PLEG): container finished" podID="cdc575e3-fc82-48e4-ae54-27f374e37063" containerID="301f13bc924830f186afba1bebfa0f509472a654adc102a985467ad83327e434" exitCode=0 Oct 03 13:53:37 crc kubenswrapper[4959]: I1003 13:53:37.300400 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cd4-account-create-6tgxr" event={"ID":"cdc575e3-fc82-48e4-ae54-27f374e37063","Type":"ContainerDied","Data":"301f13bc924830f186afba1bebfa0f509472a654adc102a985467ad83327e434"} Oct 03 13:53:37 crc kubenswrapper[4959]: I1003 13:53:37.320802 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerStarted","Data":"b304c4a31b60f0e9c502c5bfcdf56c0274ea52bcd6515a31c6e5973c7f3cb733"} Oct 03 13:53:37 crc kubenswrapper[4959]: I1003 13:53:37.342228 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerStarted","Data":"17a8c4927f623d23c41b21fe1fe20fe9b73cc2ce535689ee98c05394d443585e"} Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.354173 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerStarted","Data":"0f239875c91ee1dd657942f1b4a5bee065b5388e0803e40d082ef09005a99e55"} Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.357978 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerStarted","Data":"c00cd6709b9443f9ebdd1fe25f8f6925029bbbed2d672d187b74ef5891961905"} Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.378905 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.378886715 podStartE2EDuration="3.378886715s" podCreationTimestamp="2025-10-03 13:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:53:38.372977872 +0000 UTC m=+1387.576321289" watchObservedRunningTime="2025-10-03 13:53:38.378886715 +0000 UTC m=+1387.582230132" Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.747716 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.905493 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t8tc\" (UniqueName: \"kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc\") pod \"cdc575e3-fc82-48e4-ae54-27f374e37063\" (UID: \"cdc575e3-fc82-48e4-ae54-27f374e37063\") " Oct 03 13:53:38 crc kubenswrapper[4959]: I1003 13:53:38.916354 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc" (OuterVolumeSpecName: "kube-api-access-8t8tc") pod "cdc575e3-fc82-48e4-ae54-27f374e37063" (UID: "cdc575e3-fc82-48e4-ae54-27f374e37063"). InnerVolumeSpecName "kube-api-access-8t8tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:39 crc kubenswrapper[4959]: I1003 13:53:39.007879 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t8tc\" (UniqueName: \"kubernetes.io/projected/cdc575e3-fc82-48e4-ae54-27f374e37063-kube-api-access-8t8tc\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:39 crc kubenswrapper[4959]: I1003 13:53:39.179784 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 03 13:53:39 crc kubenswrapper[4959]: I1003 13:53:39.370750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3cd4-account-create-6tgxr" Oct 03 13:53:39 crc kubenswrapper[4959]: I1003 13:53:39.371166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3cd4-account-create-6tgxr" event={"ID":"cdc575e3-fc82-48e4-ae54-27f374e37063","Type":"ContainerDied","Data":"f50c54a3c106c5171991105a105452894a8795d9cf98f093622d0a4128570dd2"} Oct 03 13:53:39 crc kubenswrapper[4959]: I1003 13:53:39.371268 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f50c54a3c106c5171991105a105452894a8795d9cf98f093622d0a4128570dd2" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.215815 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.340734 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdt5h\" (UniqueName: \"kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h\") pod \"47d02229-11f4-4ad4-ad25-698800155fdb\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.340853 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config\") pod \"47d02229-11f4-4ad4-ad25-698800155fdb\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.340935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle\") pod \"47d02229-11f4-4ad4-ad25-698800155fdb\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.341013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config\") pod \"47d02229-11f4-4ad4-ad25-698800155fdb\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.341084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs\") pod \"47d02229-11f4-4ad4-ad25-698800155fdb\" (UID: \"47d02229-11f4-4ad4-ad25-698800155fdb\") " Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.371371 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h" (OuterVolumeSpecName: "kube-api-access-xdt5h") pod "47d02229-11f4-4ad4-ad25-698800155fdb" (UID: "47d02229-11f4-4ad4-ad25-698800155fdb"). InnerVolumeSpecName "kube-api-access-xdt5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.371373 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "47d02229-11f4-4ad4-ad25-698800155fdb" (UID: "47d02229-11f4-4ad4-ad25-698800155fdb"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.424556 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerStarted","Data":"13616c03bc68db361e7b2adb61c924083b72860218ab97ae4f0e19227a2a1f3a"} Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.424759 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-central-agent" containerID="cri-o://cef4485dd143a31defd4cf196d86695d90b172ab45acac529670496587ccfe92" gracePeriod=30 Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.425050 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.425400 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="proxy-httpd" containerID="cri-o://13616c03bc68db361e7b2adb61c924083b72860218ab97ae4f0e19227a2a1f3a" gracePeriod=30 Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.425460 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="sg-core" containerID="cri-o://c00cd6709b9443f9ebdd1fe25f8f6925029bbbed2d672d187b74ef5891961905" gracePeriod=30 Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.425502 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-notification-agent" containerID="cri-o://17a8c4927f623d23c41b21fe1fe20fe9b73cc2ce535689ee98c05394d443585e" gracePeriod=30 Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.437683 4959 generic.go:334] "Generic (PLEG): container finished" podID="47d02229-11f4-4ad4-ad25-698800155fdb" containerID="102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc" exitCode=0 Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.437733 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerDied","Data":"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc"} Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.437763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-596475bc7b-44mtq" event={"ID":"47d02229-11f4-4ad4-ad25-698800155fdb","Type":"ContainerDied","Data":"2744e49cf24e053f183c0f3a9e6ce49a3599757356b98393c127fcd8c8b972d0"} Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.437782 4959 scope.go:117] "RemoveContainer" containerID="9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.437921 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-596475bc7b-44mtq" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.459903 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdt5h\" (UniqueName: \"kubernetes.io/projected/47d02229-11f4-4ad4-ad25-698800155fdb-kube-api-access-xdt5h\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.459978 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.478618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config" (OuterVolumeSpecName: "config") pod "47d02229-11f4-4ad4-ad25-698800155fdb" (UID: "47d02229-11f4-4ad4-ad25-698800155fdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.488233 4959 scope.go:117] "RemoveContainer" containerID="102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.488228 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47d02229-11f4-4ad4-ad25-698800155fdb" (UID: "47d02229-11f4-4ad4-ad25-698800155fdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.519402 4959 scope.go:117] "RemoveContainer" containerID="9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d" Oct 03 13:53:40 crc kubenswrapper[4959]: E1003 13:53:40.519839 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d\": container with ID starting with 9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d not found: ID does not exist" containerID="9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.519885 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d"} err="failed to get container status \"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d\": rpc error: code = NotFound desc = could not find container \"9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d\": container with ID starting with 9899436af040cd8fc6b12c75764a146bc5ecac291b3b8e30c44690bfcad5e61d not found: ID does not exist" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.519909 4959 scope.go:117] "RemoveContainer" containerID="102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc" Oct 03 13:53:40 crc kubenswrapper[4959]: E1003 13:53:40.520171 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc\": container with ID starting with 102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc not found: ID does not exist" containerID="102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.520227 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc"} err="failed to get container status \"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc\": rpc error: code = NotFound desc = could not find container \"102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc\": container with ID starting with 102a71c84276c161b0a54a67c826fbd57c391a1d6d6ac887158005ac237558bc not found: ID does not exist" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.520883 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.169580093 podStartE2EDuration="6.520849724s" podCreationTimestamp="2025-10-03 13:53:34 +0000 UTC" firstStartedPulling="2025-10-03 13:53:35.130049756 +0000 UTC m=+1384.333393173" lastFinishedPulling="2025-10-03 13:53:39.481319387 +0000 UTC m=+1388.684662804" observedRunningTime="2025-10-03 13:53:40.471510967 +0000 UTC m=+1389.674854404" watchObservedRunningTime="2025-10-03 13:53:40.520849724 +0000 UTC m=+1389.724193151" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.528526 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.528570 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.529790 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "47d02229-11f4-4ad4-ad25-698800155fdb" (UID: "47d02229-11f4-4ad4-ad25-698800155fdb"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.561362 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.561405 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.561416 4959 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/47d02229-11f4-4ad4-ad25-698800155fdb-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.561630 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.572627 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.681889 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.771042 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:40 crc kubenswrapper[4959]: I1003 13:53:40.782962 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-596475bc7b-44mtq"] Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448403 4959 generic.go:334] "Generic (PLEG): container finished" podID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerID="13616c03bc68db361e7b2adb61c924083b72860218ab97ae4f0e19227a2a1f3a" exitCode=0 Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448443 4959 generic.go:334] "Generic (PLEG): container finished" podID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerID="c00cd6709b9443f9ebdd1fe25f8f6925029bbbed2d672d187b74ef5891961905" exitCode=2 Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448454 4959 generic.go:334] "Generic (PLEG): container finished" podID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerID="17a8c4927f623d23c41b21fe1fe20fe9b73cc2ce535689ee98c05394d443585e" exitCode=0 Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448496 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerDied","Data":"13616c03bc68db361e7b2adb61c924083b72860218ab97ae4f0e19227a2a1f3a"} Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerDied","Data":"c00cd6709b9443f9ebdd1fe25f8f6925029bbbed2d672d187b74ef5891961905"} Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.448539 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerDied","Data":"17a8c4927f623d23c41b21fe1fe20fe9b73cc2ce535689ee98c05394d443585e"} Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.454125 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.454157 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:53:41 crc kubenswrapper[4959]: I1003 13:53:41.696659 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" path="/var/lib/kubelet/pods/47d02229-11f4-4ad4-ad25-698800155fdb/volumes" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.499771 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.500766 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.503848 4959 generic.go:334] "Generic (PLEG): container finished" podID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerID="cef4485dd143a31defd4cf196d86695d90b172ab45acac529670496587ccfe92" exitCode=0 Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.504683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerDied","Data":"cef4485dd143a31defd4cf196d86695d90b172ab45acac529670496587ccfe92"} Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.764702 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.800709 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.800762 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823358 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmjzv\" (UniqueName: \"kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823448 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823495 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823537 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823579 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823620 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.823677 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts\") pod \"2393fe91-a4ae-41fb-b6df-e38997b9510c\" (UID: \"2393fe91-a4ae-41fb-b6df-e38997b9510c\") " Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.825175 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.826009 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.831360 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv" (OuterVolumeSpecName: "kube-api-access-bmjzv") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "kube-api-access-bmjzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.832038 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts" (OuterVolumeSpecName: "scripts") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.855538 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.856494 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.877465 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.926099 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmjzv\" (UniqueName: \"kubernetes.io/projected/2393fe91-a4ae-41fb-b6df-e38997b9510c-kube-api-access-bmjzv\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.926145 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.926158 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.926170 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2393fe91-a4ae-41fb-b6df-e38997b9510c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.926181 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.940437 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:43 crc kubenswrapper[4959]: I1003 13:53:43.968371 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data" (OuterVolumeSpecName: "config-data") pod "2393fe91-a4ae-41fb-b6df-e38997b9510c" (UID: "2393fe91-a4ae-41fb-b6df-e38997b9510c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.028734 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.028783 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2393fe91-a4ae-41fb-b6df-e38997b9510c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.515571 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2393fe91-a4ae-41fb-b6df-e38997b9510c","Type":"ContainerDied","Data":"3dd937b5c4e769e9e6ce845a3dcff4ee73580495bb9019b6917cda690b07df00"} Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.515626 4959 scope.go:117] "RemoveContainer" containerID="13616c03bc68db361e7b2adb61c924083b72860218ab97ae4f0e19227a2a1f3a" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.515956 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.516136 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.516180 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.537866 4959 scope.go:117] "RemoveContainer" containerID="c00cd6709b9443f9ebdd1fe25f8f6925029bbbed2d672d187b74ef5891961905" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.555312 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.578261 4959 scope.go:117] "RemoveContainer" containerID="17a8c4927f623d23c41b21fe1fe20fe9b73cc2ce535689ee98c05394d443585e" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.590605 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.601473 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602166 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-api" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602186 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-api" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602220 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="proxy-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602228 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="proxy-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602238 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc575e3-fc82-48e4-ae54-27f374e37063" containerName="mariadb-account-create" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602246 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc575e3-fc82-48e4-ae54-27f374e37063" containerName="mariadb-account-create" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602265 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602272 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602283 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="sg-core" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602289 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="sg-core" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602301 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-central-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602307 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-central-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: E1003 13:53:44.602320 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-notification-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602327 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-notification-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602526 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc575e3-fc82-48e4-ae54-27f374e37063" containerName="mariadb-account-create" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602538 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="proxy-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602553 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-notification-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602566 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="sg-core" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602577 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" containerName="ceilometer-central-agent" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602591 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-api" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.602606 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d02229-11f4-4ad4-ad25-698800155fdb" containerName="neutron-httpd" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.608465 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.612584 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.613045 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.613267 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.632321 4959 scope.go:117] "RemoveContainer" containerID="cef4485dd143a31defd4cf196d86695d90b172ab45acac529670496587ccfe92" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742607 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742672 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5v2j\" (UniqueName: \"kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742774 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742843 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.742919 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844755 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844819 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844841 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844870 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844893 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.844908 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5v2j\" (UniqueName: \"kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.845532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.846128 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.854697 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.854972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.902787 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.905102 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.919468 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5v2j\" (UniqueName: \"kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j\") pod \"ceilometer-0\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " pod="openstack/ceilometer-0" Oct 03 13:53:44 crc kubenswrapper[4959]: I1003 13:53:44.932833 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.098392 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="779685e0-3882-4dbe-ad13-df648b89d1a0" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.142:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.294550 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ef0b-account-create-4xq88"] Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.295666 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.298121 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.305684 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef0b-account-create-4xq88"] Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.355316 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j64fw\" (UniqueName: \"kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw\") pod \"nova-cell0-ef0b-account-create-4xq88\" (UID: \"d4c3386b-7ed5-4876-bbcb-f00137430c8b\") " pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.440822 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.456877 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j64fw\" (UniqueName: \"kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw\") pod \"nova-cell0-ef0b-account-create-4xq88\" (UID: \"d4c3386b-7ed5-4876-bbcb-f00137430c8b\") " pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.481658 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j64fw\" (UniqueName: \"kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw\") pod \"nova-cell0-ef0b-account-create-4xq88\" (UID: \"d4c3386b-7ed5-4876-bbcb-f00137430c8b\") " pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.499140 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8c4f-account-create-mhwmc"] Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.513778 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8c4f-account-create-mhwmc"] Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.513877 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.517049 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.527561 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerStarted","Data":"7d6ce2958ed8fef4fb12704b298fc10f6e4c61691d46378cb617129000f7a004"} Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.560387 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfqsn\" (UniqueName: \"kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn\") pod \"nova-cell1-8c4f-account-create-mhwmc\" (UID: \"6ae02254-e00c-447a-84a9-4ab1b433b5e8\") " pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.624295 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.662311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfqsn\" (UniqueName: \"kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn\") pod \"nova-cell1-8c4f-account-create-mhwmc\" (UID: \"6ae02254-e00c-447a-84a9-4ab1b433b5e8\") " pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.684674 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfqsn\" (UniqueName: \"kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn\") pod \"nova-cell1-8c4f-account-create-mhwmc\" (UID: \"6ae02254-e00c-447a-84a9-4ab1b433b5e8\") " pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.701488 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2393fe91-a4ae-41fb-b6df-e38997b9510c" path="/var/lib/kubelet/pods/2393fe91-a4ae-41fb-b6df-e38997b9510c/volumes" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.858490 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:45 crc kubenswrapper[4959]: I1003 13:53:45.990736 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.119929 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef0b-account-create-4xq88"] Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.327165 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8c4f-account-create-mhwmc"] Oct 03 13:53:46 crc kubenswrapper[4959]: W1003 13:53:46.340220 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ae02254_e00c_447a_84a9_4ab1b433b5e8.slice/crio-b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1 WatchSource:0}: Error finding container b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1: Status 404 returned error can't find the container with id b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1 Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.542069 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerStarted","Data":"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025"} Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.543371 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" event={"ID":"6ae02254-e00c-447a-84a9-4ab1b433b5e8","Type":"ContainerStarted","Data":"b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1"} Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.544741 4959 generic.go:334] "Generic (PLEG): container finished" podID="d4c3386b-7ed5-4876-bbcb-f00137430c8b" containerID="a245794e4104503620a1e62233334902247449b3325e793e6c53eee6dfed9e0a" exitCode=0 Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.544777 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef0b-account-create-4xq88" event={"ID":"d4c3386b-7ed5-4876-bbcb-f00137430c8b","Type":"ContainerDied","Data":"a245794e4104503620a1e62233334902247449b3325e793e6c53eee6dfed9e0a"} Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.544798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef0b-account-create-4xq88" event={"ID":"d4c3386b-7ed5-4876-bbcb-f00137430c8b","Type":"ContainerStarted","Data":"cd5620cf27a7f5ef10137ed4993ed92c95b998aa56ec784cf3c59d172a0487b8"} Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.769167 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.769335 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:53:46 crc kubenswrapper[4959]: I1003 13:53:46.817564 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:53:47 crc kubenswrapper[4959]: I1003 13:53:47.556052 4959 generic.go:334] "Generic (PLEG): container finished" podID="6ae02254-e00c-447a-84a9-4ab1b433b5e8" containerID="d8be0657675a6b800321dc24a5acf47e262aa26ce6657e72b068da8f72f7bdef" exitCode=0 Oct 03 13:53:47 crc kubenswrapper[4959]: I1003 13:53:47.556111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" event={"ID":"6ae02254-e00c-447a-84a9-4ab1b433b5e8","Type":"ContainerDied","Data":"d8be0657675a6b800321dc24a5acf47e262aa26ce6657e72b068da8f72f7bdef"} Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.003961 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.107769 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j64fw\" (UniqueName: \"kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw\") pod \"d4c3386b-7ed5-4876-bbcb-f00137430c8b\" (UID: \"d4c3386b-7ed5-4876-bbcb-f00137430c8b\") " Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.112593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw" (OuterVolumeSpecName: "kube-api-access-j64fw") pod "d4c3386b-7ed5-4876-bbcb-f00137430c8b" (UID: "d4c3386b-7ed5-4876-bbcb-f00137430c8b"). InnerVolumeSpecName "kube-api-access-j64fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.209014 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j64fw\" (UniqueName: \"kubernetes.io/projected/d4c3386b-7ed5-4876-bbcb-f00137430c8b-kube-api-access-j64fw\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.566248 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerStarted","Data":"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d"} Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.568236 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef0b-account-create-4xq88" event={"ID":"d4c3386b-7ed5-4876-bbcb-f00137430c8b","Type":"ContainerDied","Data":"cd5620cf27a7f5ef10137ed4993ed92c95b998aa56ec784cf3c59d172a0487b8"} Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.568277 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd5620cf27a7f5ef10137ed4993ed92c95b998aa56ec784cf3c59d172a0487b8" Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.568276 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef0b-account-create-4xq88" Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.847083 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:48 crc kubenswrapper[4959]: I1003 13:53:48.940152 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.136440 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfqsn\" (UniqueName: \"kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn\") pod \"6ae02254-e00c-447a-84a9-4ab1b433b5e8\" (UID: \"6ae02254-e00c-447a-84a9-4ab1b433b5e8\") " Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.142700 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn" (OuterVolumeSpecName: "kube-api-access-rfqsn") pod "6ae02254-e00c-447a-84a9-4ab1b433b5e8" (UID: "6ae02254-e00c-447a-84a9-4ab1b433b5e8"). InnerVolumeSpecName "kube-api-access-rfqsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.238877 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfqsn\" (UniqueName: \"kubernetes.io/projected/6ae02254-e00c-447a-84a9-4ab1b433b5e8-kube-api-access-rfqsn\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.578352 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerStarted","Data":"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7"} Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.580721 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" event={"ID":"6ae02254-e00c-447a-84a9-4ab1b433b5e8","Type":"ContainerDied","Data":"b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1"} Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.580836 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4f1df56745314af544b6c028055452feba67ac33570eeea64068f9cb2b22ca1" Oct 03 13:53:49 crc kubenswrapper[4959]: I1003 13:53:49.580925 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8c4f-account-create-mhwmc" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.461450 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6pgdx"] Oct 03 13:53:50 crc kubenswrapper[4959]: E1003 13:53:50.462080 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae02254-e00c-447a-84a9-4ab1b433b5e8" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.462099 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae02254-e00c-447a-84a9-4ab1b433b5e8" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: E1003 13:53:50.462125 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c3386b-7ed5-4876-bbcb-f00137430c8b" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.462136 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c3386b-7ed5-4876-bbcb-f00137430c8b" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.462367 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae02254-e00c-447a-84a9-4ab1b433b5e8" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.462398 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c3386b-7ed5-4876-bbcb-f00137430c8b" containerName="mariadb-account-create" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.463076 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.470523 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.471664 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sjklq" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.472438 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.481548 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6pgdx"] Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.559428 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.559710 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.559837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m8xq\" (UniqueName: \"kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.559959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.661429 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m8xq\" (UniqueName: \"kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.661523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.661565 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.661616 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.665934 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.666708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.674966 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.682922 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m8xq\" (UniqueName: \"kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq\") pod \"nova-cell0-conductor-db-sync-6pgdx\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:50 crc kubenswrapper[4959]: I1003 13:53:50.786518 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.231581 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6pgdx"] Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.598518 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" event={"ID":"ceeec8b6-407f-4a35-a745-d51f676bcbaf","Type":"ContainerStarted","Data":"ddda429364948d93fd367017593e6b27984ecb83a54e6e24989db7a5743018d6"} Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerStarted","Data":"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277"} Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601467 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-central-agent" containerID="cri-o://9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025" gracePeriod=30 Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601506 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="proxy-httpd" containerID="cri-o://d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277" gracePeriod=30 Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601542 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-notification-agent" containerID="cri-o://a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d" gracePeriod=30 Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601483 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.601559 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="sg-core" containerID="cri-o://fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7" gracePeriod=30 Oct 03 13:53:51 crc kubenswrapper[4959]: I1003 13:53:51.625798 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.390754079 podStartE2EDuration="7.625780786s" podCreationTimestamp="2025-10-03 13:53:44 +0000 UTC" firstStartedPulling="2025-10-03 13:53:45.449143829 +0000 UTC m=+1394.652487236" lastFinishedPulling="2025-10-03 13:53:50.684170526 +0000 UTC m=+1399.887513943" observedRunningTime="2025-10-03 13:53:51.621505272 +0000 UTC m=+1400.824848689" watchObservedRunningTime="2025-10-03 13:53:51.625780786 +0000 UTC m=+1400.829124203" Oct 03 13:53:51 crc kubenswrapper[4959]: E1003 13:53:51.867091 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa0ae31b_09ae_41aa_82d5_e17581433d0e.slice/crio-d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618863 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerID="d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277" exitCode=0 Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618909 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerID="fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7" exitCode=2 Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618920 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerID="a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d" exitCode=0 Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618919 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerDied","Data":"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277"} Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerDied","Data":"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7"} Oct 03 13:53:52 crc kubenswrapper[4959]: I1003 13:53:52.618988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerDied","Data":"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d"} Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.079607 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127133 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5v2j\" (UniqueName: \"kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127177 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127225 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127294 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127335 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127364 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.127466 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts\") pod \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\" (UID: \"aa0ae31b-09ae-41aa-82d5-e17581433d0e\") " Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.128553 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.128568 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.148370 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts" (OuterVolumeSpecName: "scripts") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.152051 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j" (OuterVolumeSpecName: "kube-api-access-f5v2j") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "kube-api-access-f5v2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.159624 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.224860 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229714 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229743 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229808 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5v2j\" (UniqueName: \"kubernetes.io/projected/aa0ae31b-09ae-41aa-82d5-e17581433d0e-kube-api-access-f5v2j\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229824 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229835 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa0ae31b-09ae-41aa-82d5-e17581433d0e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.229846 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.239789 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data" (OuterVolumeSpecName: "config-data") pod "aa0ae31b-09ae-41aa-82d5-e17581433d0e" (UID: "aa0ae31b-09ae-41aa-82d5-e17581433d0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.331243 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa0ae31b-09ae-41aa-82d5-e17581433d0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.638985 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerID="9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025" exitCode=0 Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.639038 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerDied","Data":"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025"} Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.639046 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.639074 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa0ae31b-09ae-41aa-82d5-e17581433d0e","Type":"ContainerDied","Data":"7d6ce2958ed8fef4fb12704b298fc10f6e4c61691d46378cb617129000f7a004"} Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.639096 4959 scope.go:117] "RemoveContainer" containerID="d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.672461 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.682097 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.703739 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:54 crc kubenswrapper[4959]: E1003 13:53:54.704206 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-central-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704226 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-central-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: E1003 13:53:54.704253 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-notification-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704262 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-notification-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: E1003 13:53:54.704281 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="proxy-httpd" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704288 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="proxy-httpd" Oct 03 13:53:54 crc kubenswrapper[4959]: E1003 13:53:54.704301 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="sg-core" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704309 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="sg-core" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704516 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-notification-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704544 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="ceilometer-central-agent" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704561 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="proxy-httpd" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.704575 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" containerName="sg-core" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.722409 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.726440 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.726488 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.728851 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738437 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738481 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738518 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738578 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738615 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.738657 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.741993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2h68\" (UniqueName: \"kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845149 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845273 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845298 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2h68\" (UniqueName: \"kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845759 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845799 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.845819 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.852152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.852772 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.855098 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.855320 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.856346 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.863316 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2h68\" (UniqueName: \"kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68\") pod \"ceilometer-0\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " pod="openstack/ceilometer-0" Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.868466 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:53:54 crc kubenswrapper[4959]: I1003 13:53:54.869141 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:53:55 crc kubenswrapper[4959]: I1003 13:53:55.696348 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa0ae31b-09ae-41aa-82d5-e17581433d0e" path="/var/lib/kubelet/pods/aa0ae31b-09ae-41aa-82d5-e17581433d0e/volumes" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.195456 4959 scope.go:117] "RemoveContainer" containerID="fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.288792 4959 scope.go:117] "RemoveContainer" containerID="a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.437370 4959 scope.go:117] "RemoveContainer" containerID="9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.468927 4959 scope.go:117] "RemoveContainer" containerID="d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277" Oct 03 13:53:59 crc kubenswrapper[4959]: E1003 13:53:59.469754 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277\": container with ID starting with d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277 not found: ID does not exist" containerID="d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.469821 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277"} err="failed to get container status \"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277\": rpc error: code = NotFound desc = could not find container \"d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277\": container with ID starting with d118603b8c40d1ee8aa193179ca911a33e5c32523b9323c1793bea8e802ad277 not found: ID does not exist" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.469858 4959 scope.go:117] "RemoveContainer" containerID="fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7" Oct 03 13:53:59 crc kubenswrapper[4959]: E1003 13:53:59.470517 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7\": container with ID starting with fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7 not found: ID does not exist" containerID="fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.470569 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7"} err="failed to get container status \"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7\": rpc error: code = NotFound desc = could not find container \"fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7\": container with ID starting with fbe6d3def6e79c6e6bf25f89aae09324c34a26e833e4d806979c4f467ab71db7 not found: ID does not exist" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.470600 4959 scope.go:117] "RemoveContainer" containerID="a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d" Oct 03 13:53:59 crc kubenswrapper[4959]: E1003 13:53:59.471086 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d\": container with ID starting with a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d not found: ID does not exist" containerID="a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.471123 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d"} err="failed to get container status \"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d\": rpc error: code = NotFound desc = could not find container \"a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d\": container with ID starting with a9b1aaff08800e0adb6d346bd530628e65b846b40d495fa68b666e5722a0814d not found: ID does not exist" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.471144 4959 scope.go:117] "RemoveContainer" containerID="9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025" Oct 03 13:53:59 crc kubenswrapper[4959]: E1003 13:53:59.471480 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025\": container with ID starting with 9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025 not found: ID does not exist" containerID="9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025" Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.471529 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025"} err="failed to get container status \"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025\": rpc error: code = NotFound desc = could not find container \"9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025\": container with ID starting with 9846dcf8de54516db9e1ce29e4b659129d26efcaf21f56d8af8d3d2d107e4025 not found: ID does not exist" Oct 03 13:53:59 crc kubenswrapper[4959]: W1003 13:53:59.738646 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode20e1301_2436_4aed_b47b_850be2af39a2.slice/crio-f3d1ddddc7f24d5bd9d053edc4335c889d9c5bd996a876dd40af522583afadcb WatchSource:0}: Error finding container f3d1ddddc7f24d5bd9d053edc4335c889d9c5bd996a876dd40af522583afadcb: Status 404 returned error can't find the container with id f3d1ddddc7f24d5bd9d053edc4335c889d9c5bd996a876dd40af522583afadcb Oct 03 13:53:59 crc kubenswrapper[4959]: I1003 13:53:59.740892 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:00 crc kubenswrapper[4959]: I1003 13:54:00.701807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" event={"ID":"ceeec8b6-407f-4a35-a745-d51f676bcbaf","Type":"ContainerStarted","Data":"6b617a3bf18abe83bfc72274e793ce4a300223427e3b278bb1ccc7c7bec09bcc"} Oct 03 13:54:00 crc kubenswrapper[4959]: I1003 13:54:00.705958 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerStarted","Data":"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c"} Oct 03 13:54:00 crc kubenswrapper[4959]: I1003 13:54:00.706146 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerStarted","Data":"f3d1ddddc7f24d5bd9d053edc4335c889d9c5bd996a876dd40af522583afadcb"} Oct 03 13:54:00 crc kubenswrapper[4959]: I1003 13:54:00.721122 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" podStartSLOduration=2.64012926 podStartE2EDuration="10.721103751s" podCreationTimestamp="2025-10-03 13:53:50 +0000 UTC" firstStartedPulling="2025-10-03 13:53:51.239141423 +0000 UTC m=+1400.442484840" lastFinishedPulling="2025-10-03 13:53:59.320115914 +0000 UTC m=+1408.523459331" observedRunningTime="2025-10-03 13:54:00.720103467 +0000 UTC m=+1409.923446894" watchObservedRunningTime="2025-10-03 13:54:00.721103751 +0000 UTC m=+1409.924447168" Oct 03 13:54:05 crc kubenswrapper[4959]: I1003 13:54:05.757177 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerStarted","Data":"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd"} Oct 03 13:54:12 crc kubenswrapper[4959]: I1003 13:54:12.834935 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerStarted","Data":"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62"} Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.844848 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-central-agent" containerID="cri-o://8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c" gracePeriod=30 Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.844885 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="proxy-httpd" containerID="cri-o://2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac" gracePeriod=30 Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.844946 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-notification-agent" containerID="cri-o://a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd" gracePeriod=30 Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.844669 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerStarted","Data":"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac"} Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.845387 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.847816 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="sg-core" containerID="cri-o://af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62" gracePeriod=30 Oct 03 13:54:13 crc kubenswrapper[4959]: I1003 13:54:13.876208 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.441534838 podStartE2EDuration="19.876174235s" podCreationTimestamp="2025-10-03 13:53:54 +0000 UTC" firstStartedPulling="2025-10-03 13:53:59.740805773 +0000 UTC m=+1408.944149190" lastFinishedPulling="2025-10-03 13:54:13.17544516 +0000 UTC m=+1422.378788587" observedRunningTime="2025-10-03 13:54:13.865338382 +0000 UTC m=+1423.068681809" watchObservedRunningTime="2025-10-03 13:54:13.876174235 +0000 UTC m=+1423.079517642" Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.854999 4959 generic.go:334] "Generic (PLEG): container finished" podID="e20e1301-2436-4aed-b47b-850be2af39a2" containerID="2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac" exitCode=0 Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.855031 4959 generic.go:334] "Generic (PLEG): container finished" podID="e20e1301-2436-4aed-b47b-850be2af39a2" containerID="af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62" exitCode=2 Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.855038 4959 generic.go:334] "Generic (PLEG): container finished" podID="e20e1301-2436-4aed-b47b-850be2af39a2" containerID="a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd" exitCode=0 Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.855038 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerDied","Data":"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac"} Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.855084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerDied","Data":"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62"} Oct 03 13:54:14 crc kubenswrapper[4959]: I1003 13:54:14.855099 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerDied","Data":"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd"} Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.720159 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775392 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775505 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775614 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775667 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2h68\" (UniqueName: \"kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775768 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.775816 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd\") pod \"e20e1301-2436-4aed-b47b-850be2af39a2\" (UID: \"e20e1301-2436-4aed-b47b-850be2af39a2\") " Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.776568 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.776771 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.784248 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts" (OuterVolumeSpecName: "scripts") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.804413 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68" (OuterVolumeSpecName: "kube-api-access-s2h68") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "kube-api-access-s2h68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.812254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.870986 4959 generic.go:334] "Generic (PLEG): container finished" podID="e20e1301-2436-4aed-b47b-850be2af39a2" containerID="8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c" exitCode=0 Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.871040 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerDied","Data":"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c"} Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.871095 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e20e1301-2436-4aed-b47b-850be2af39a2","Type":"ContainerDied","Data":"f3d1ddddc7f24d5bd9d053edc4335c889d9c5bd996a876dd40af522583afadcb"} Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.871119 4959 scope.go:117] "RemoveContainer" containerID="2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.871142 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.874861 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878043 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878070 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878084 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2h68\" (UniqueName: \"kubernetes.io/projected/e20e1301-2436-4aed-b47b-850be2af39a2-kube-api-access-s2h68\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878097 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878110 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e20e1301-2436-4aed-b47b-850be2af39a2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878121 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.878252 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data" (OuterVolumeSpecName: "config-data") pod "e20e1301-2436-4aed-b47b-850be2af39a2" (UID: "e20e1301-2436-4aed-b47b-850be2af39a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.948340 4959 scope.go:117] "RemoveContainer" containerID="af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.974242 4959 scope.go:117] "RemoveContainer" containerID="a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.979927 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1301-2436-4aed-b47b-850be2af39a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:15 crc kubenswrapper[4959]: I1003 13:54:15.997947 4959 scope.go:117] "RemoveContainer" containerID="8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.020577 4959 scope.go:117] "RemoveContainer" containerID="2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.020989 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac\": container with ID starting with 2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac not found: ID does not exist" containerID="2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.021047 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac"} err="failed to get container status \"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac\": rpc error: code = NotFound desc = could not find container \"2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac\": container with ID starting with 2ac1b3cccfb5906121a3a69ae8d3e7605af4fe0ccf2f7556724252be4f5e42ac not found: ID does not exist" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.021078 4959 scope.go:117] "RemoveContainer" containerID="af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.021468 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62\": container with ID starting with af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62 not found: ID does not exist" containerID="af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.021498 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62"} err="failed to get container status \"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62\": rpc error: code = NotFound desc = could not find container \"af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62\": container with ID starting with af2cb6c1e4b46c57cd639420a06ee8969ebb70e88740ac6f57fa463b48e12d62 not found: ID does not exist" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.021516 4959 scope.go:117] "RemoveContainer" containerID="a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.021983 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd\": container with ID starting with a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd not found: ID does not exist" containerID="a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.022029 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd"} err="failed to get container status \"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd\": rpc error: code = NotFound desc = could not find container \"a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd\": container with ID starting with a414d72647d167d815021bec221289fe3aef0854a0698aea2dda65ff0d5ca4dd not found: ID does not exist" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.022061 4959 scope.go:117] "RemoveContainer" containerID="8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.022355 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c\": container with ID starting with 8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c not found: ID does not exist" containerID="8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.022389 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c"} err="failed to get container status \"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c\": rpc error: code = NotFound desc = could not find container \"8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c\": container with ID starting with 8b65b48cabc2873df7d5fa75175e1437e8641cc23bca6758c4f2711209d4761c not found: ID does not exist" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.213113 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.220137 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.243911 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.244363 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="proxy-httpd" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244386 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="proxy-httpd" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.244404 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="sg-core" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244415 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="sg-core" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.244460 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-central-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244468 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-central-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: E1003 13:54:16.244480 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-notification-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244489 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-notification-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244703 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="sg-core" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244728 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-central-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.244740 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="ceilometer-notification-agent" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.246787 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" containerName="proxy-httpd" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.248760 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.252578 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.252808 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.253951 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285415 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285482 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285539 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285589 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x5wg\" (UniqueName: \"kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285647 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285676 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.285693 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x5wg\" (UniqueName: \"kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.386758 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.387118 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.387173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.390338 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.390380 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.390528 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.390947 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.402980 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x5wg\" (UniqueName: \"kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg\") pod \"ceilometer-0\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " pod="openstack/ceilometer-0" Oct 03 13:54:16 crc kubenswrapper[4959]: I1003 13:54:16.576028 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:17 crc kubenswrapper[4959]: I1003 13:54:17.037975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:17 crc kubenswrapper[4959]: I1003 13:54:17.696557 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e20e1301-2436-4aed-b47b-850be2af39a2" path="/var/lib/kubelet/pods/e20e1301-2436-4aed-b47b-850be2af39a2/volumes" Oct 03 13:54:17 crc kubenswrapper[4959]: I1003 13:54:17.891287 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerStarted","Data":"3b2ab7b13434044306549d54bc915d4778e56be24c7855da96228c6a69b18300"} Oct 03 13:54:17 crc kubenswrapper[4959]: I1003 13:54:17.891571 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerStarted","Data":"22d9ec904cb11528b29cfa20d9b5ba6a378bb62076311a3d7e3fa18bd4947741"} Oct 03 13:54:18 crc kubenswrapper[4959]: I1003 13:54:18.903109 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerStarted","Data":"5bb3fa231f29ee6dab17213f0ac78b3f8a0b41725add6f76cf68a89d2897b04e"} Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.747269 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.749154 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.777246 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.847153 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.847331 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.847402 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtjgf\" (UniqueName: \"kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.913680 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerStarted","Data":"5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c"} Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.949299 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtjgf\" (UniqueName: \"kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.949409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.949526 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.950039 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.950098 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:19 crc kubenswrapper[4959]: I1003 13:54:19.971916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtjgf\" (UniqueName: \"kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf\") pod \"certified-operators-j7whn\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:20 crc kubenswrapper[4959]: I1003 13:54:20.067819 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:20 crc kubenswrapper[4959]: W1003 13:54:20.609362 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd29d59a_9ff4_4300_b193_84e5b1a72274.slice/crio-14b45571d629d3909d76cef709d67ab7aa78aec71a15203cdd90b6f1ca2e7ceb WatchSource:0}: Error finding container 14b45571d629d3909d76cef709d67ab7aa78aec71a15203cdd90b6f1ca2e7ceb: Status 404 returned error can't find the container with id 14b45571d629d3909d76cef709d67ab7aa78aec71a15203cdd90b6f1ca2e7ceb Oct 03 13:54:20 crc kubenswrapper[4959]: I1003 13:54:20.616063 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:20 crc kubenswrapper[4959]: I1003 13:54:20.928665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerStarted","Data":"14b45571d629d3909d76cef709d67ab7aa78aec71a15203cdd90b6f1ca2e7ceb"} Oct 03 13:54:21 crc kubenswrapper[4959]: I1003 13:54:21.955451 4959 generic.go:334] "Generic (PLEG): container finished" podID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerID="9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e" exitCode=0 Oct 03 13:54:21 crc kubenswrapper[4959]: I1003 13:54:21.955488 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerDied","Data":"9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e"} Oct 03 13:54:22 crc kubenswrapper[4959]: I1003 13:54:22.965359 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerStarted","Data":"5cd4b2ad25a3ac5039c4b68acd1791fc1fc99a9b43b72b53b276f124d7531433"} Oct 03 13:54:23 crc kubenswrapper[4959]: I1003 13:54:23.976566 4959 generic.go:334] "Generic (PLEG): container finished" podID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerID="2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50" exitCode=0 Oct 03 13:54:23 crc kubenswrapper[4959]: I1003 13:54:23.976674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerDied","Data":"2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50"} Oct 03 13:54:23 crc kubenswrapper[4959]: I1003 13:54:23.977873 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:54:23 crc kubenswrapper[4959]: I1003 13:54:23.980347 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:54:24 crc kubenswrapper[4959]: I1003 13:54:24.026492 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.477352981 podStartE2EDuration="8.026469701s" podCreationTimestamp="2025-10-03 13:54:16 +0000 UTC" firstStartedPulling="2025-10-03 13:54:17.048605991 +0000 UTC m=+1426.251949408" lastFinishedPulling="2025-10-03 13:54:22.597722711 +0000 UTC m=+1431.801066128" observedRunningTime="2025-10-03 13:54:24.024978155 +0000 UTC m=+1433.228321602" watchObservedRunningTime="2025-10-03 13:54:24.026469701 +0000 UTC m=+1433.229813138" Oct 03 13:54:26 crc kubenswrapper[4959]: I1003 13:54:26.002768 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerStarted","Data":"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a"} Oct 03 13:54:26 crc kubenswrapper[4959]: I1003 13:54:26.022688 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j7whn" podStartSLOduration=3.360101104 podStartE2EDuration="7.022670183s" podCreationTimestamp="2025-10-03 13:54:19 +0000 UTC" firstStartedPulling="2025-10-03 13:54:22.08793537 +0000 UTC m=+1431.291278827" lastFinishedPulling="2025-10-03 13:54:25.750504489 +0000 UTC m=+1434.953847906" observedRunningTime="2025-10-03 13:54:26.017842186 +0000 UTC m=+1435.221185623" watchObservedRunningTime="2025-10-03 13:54:26.022670183 +0000 UTC m=+1435.226013600" Oct 03 13:54:30 crc kubenswrapper[4959]: I1003 13:54:30.068073 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:30 crc kubenswrapper[4959]: I1003 13:54:30.068506 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:30 crc kubenswrapper[4959]: I1003 13:54:30.113261 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:31 crc kubenswrapper[4959]: I1003 13:54:31.066612 4959 generic.go:334] "Generic (PLEG): container finished" podID="ceeec8b6-407f-4a35-a745-d51f676bcbaf" containerID="6b617a3bf18abe83bfc72274e793ce4a300223427e3b278bb1ccc7c7bec09bcc" exitCode=0 Oct 03 13:54:31 crc kubenswrapper[4959]: I1003 13:54:31.066722 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" event={"ID":"ceeec8b6-407f-4a35-a745-d51f676bcbaf","Type":"ContainerDied","Data":"6b617a3bf18abe83bfc72274e793ce4a300223427e3b278bb1ccc7c7bec09bcc"} Oct 03 13:54:31 crc kubenswrapper[4959]: I1003 13:54:31.156277 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:31 crc kubenswrapper[4959]: I1003 13:54:31.223762 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.479524 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.623056 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle\") pod \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.623400 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data\") pod \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.623479 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts\") pod \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.623538 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m8xq\" (UniqueName: \"kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq\") pod \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\" (UID: \"ceeec8b6-407f-4a35-a745-d51f676bcbaf\") " Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.630535 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq" (OuterVolumeSpecName: "kube-api-access-7m8xq") pod "ceeec8b6-407f-4a35-a745-d51f676bcbaf" (UID: "ceeec8b6-407f-4a35-a745-d51f676bcbaf"). InnerVolumeSpecName "kube-api-access-7m8xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.630569 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts" (OuterVolumeSpecName: "scripts") pod "ceeec8b6-407f-4a35-a745-d51f676bcbaf" (UID: "ceeec8b6-407f-4a35-a745-d51f676bcbaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.658567 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data" (OuterVolumeSpecName: "config-data") pod "ceeec8b6-407f-4a35-a745-d51f676bcbaf" (UID: "ceeec8b6-407f-4a35-a745-d51f676bcbaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.660174 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ceeec8b6-407f-4a35-a745-d51f676bcbaf" (UID: "ceeec8b6-407f-4a35-a745-d51f676bcbaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.725276 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.725315 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.725328 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ceeec8b6-407f-4a35-a745-d51f676bcbaf-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:32 crc kubenswrapper[4959]: I1003 13:54:32.725366 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m8xq\" (UniqueName: \"kubernetes.io/projected/ceeec8b6-407f-4a35-a745-d51f676bcbaf-kube-api-access-7m8xq\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.111785 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" event={"ID":"ceeec8b6-407f-4a35-a745-d51f676bcbaf","Type":"ContainerDied","Data":"ddda429364948d93fd367017593e6b27984ecb83a54e6e24989db7a5743018d6"} Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.111811 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6pgdx" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.111833 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddda429364948d93fd367017593e6b27984ecb83a54e6e24989db7a5743018d6" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.112610 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j7whn" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="registry-server" containerID="cri-o://95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a" gracePeriod=2 Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.251511 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:54:33 crc kubenswrapper[4959]: E1003 13:54:33.252484 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceeec8b6-407f-4a35-a745-d51f676bcbaf" containerName="nova-cell0-conductor-db-sync" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.252517 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceeec8b6-407f-4a35-a745-d51f676bcbaf" containerName="nova-cell0-conductor-db-sync" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.252903 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceeec8b6-407f-4a35-a745-d51f676bcbaf" containerName="nova-cell0-conductor-db-sync" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.253874 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.256785 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.258295 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-sjklq" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.278066 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.348334 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.348383 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.348457 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9qp4\" (UniqueName: \"kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.450448 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.450504 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.450558 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9qp4\" (UniqueName: \"kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.457474 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.464377 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.467038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9qp4\" (UniqueName: \"kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4\") pod \"nova-cell0-conductor-0\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.599920 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.605836 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.653736 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtjgf\" (UniqueName: \"kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf\") pod \"fd29d59a-9ff4-4300-b193-84e5b1a72274\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.653881 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities\") pod \"fd29d59a-9ff4-4300-b193-84e5b1a72274\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.653957 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content\") pod \"fd29d59a-9ff4-4300-b193-84e5b1a72274\" (UID: \"fd29d59a-9ff4-4300-b193-84e5b1a72274\") " Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.655556 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities" (OuterVolumeSpecName: "utilities") pod "fd29d59a-9ff4-4300-b193-84e5b1a72274" (UID: "fd29d59a-9ff4-4300-b193-84e5b1a72274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.660918 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf" (OuterVolumeSpecName: "kube-api-access-mtjgf") pod "fd29d59a-9ff4-4300-b193-84e5b1a72274" (UID: "fd29d59a-9ff4-4300-b193-84e5b1a72274"). InnerVolumeSpecName "kube-api-access-mtjgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.722236 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd29d59a-9ff4-4300-b193-84e5b1a72274" (UID: "fd29d59a-9ff4-4300-b193-84e5b1a72274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.775738 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtjgf\" (UniqueName: \"kubernetes.io/projected/fd29d59a-9ff4-4300-b193-84e5b1a72274-kube-api-access-mtjgf\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.776540 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:33 crc kubenswrapper[4959]: I1003 13:54:33.776553 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd29d59a-9ff4-4300-b193-84e5b1a72274-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.113325 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:54:34 crc kubenswrapper[4959]: W1003 13:54:34.115982 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod774fd066_a8d1_449a_8e33_37052dbb1627.slice/crio-062a5da45e7aee48cc0feaa8f34c452c68d4bd35bcb03eab23fbb3077b5620f6 WatchSource:0}: Error finding container 062a5da45e7aee48cc0feaa8f34c452c68d4bd35bcb03eab23fbb3077b5620f6: Status 404 returned error can't find the container with id 062a5da45e7aee48cc0feaa8f34c452c68d4bd35bcb03eab23fbb3077b5620f6 Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.127675 4959 generic.go:334] "Generic (PLEG): container finished" podID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerID="95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a" exitCode=0 Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.127738 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerDied","Data":"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a"} Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.127764 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7whn" event={"ID":"fd29d59a-9ff4-4300-b193-84e5b1a72274","Type":"ContainerDied","Data":"14b45571d629d3909d76cef709d67ab7aa78aec71a15203cdd90b6f1ca2e7ceb"} Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.127782 4959 scope.go:117] "RemoveContainer" containerID="95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.127873 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7whn" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.153436 4959 scope.go:117] "RemoveContainer" containerID="2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.179607 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.191230 4959 scope.go:117] "RemoveContainer" containerID="9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.191933 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j7whn"] Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.218437 4959 scope.go:117] "RemoveContainer" containerID="95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a" Oct 03 13:54:34 crc kubenswrapper[4959]: E1003 13:54:34.218960 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a\": container with ID starting with 95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a not found: ID does not exist" containerID="95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.219007 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a"} err="failed to get container status \"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a\": rpc error: code = NotFound desc = could not find container \"95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a\": container with ID starting with 95d5cf5d78d0a8ba09e2ca55247e3e858b77cc4bb4cf4086e46766be954cc61a not found: ID does not exist" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.219036 4959 scope.go:117] "RemoveContainer" containerID="2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50" Oct 03 13:54:34 crc kubenswrapper[4959]: E1003 13:54:34.219406 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50\": container with ID starting with 2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50 not found: ID does not exist" containerID="2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.219422 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50"} err="failed to get container status \"2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50\": rpc error: code = NotFound desc = could not find container \"2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50\": container with ID starting with 2b63485232de1be603bb588cd35d709b09d1853a188cd21728270ebda8e16b50 not found: ID does not exist" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.219437 4959 scope.go:117] "RemoveContainer" containerID="9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e" Oct 03 13:54:34 crc kubenswrapper[4959]: E1003 13:54:34.219793 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e\": container with ID starting with 9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e not found: ID does not exist" containerID="9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e" Oct 03 13:54:34 crc kubenswrapper[4959]: I1003 13:54:34.219812 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e"} err="failed to get container status \"9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e\": rpc error: code = NotFound desc = could not find container \"9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e\": container with ID starting with 9a59b10fbd2e713f26bdd358c4ab241c82f5bc371a1fe8ee401e15d06f13a07e not found: ID does not exist" Oct 03 13:54:35 crc kubenswrapper[4959]: I1003 13:54:35.141674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"774fd066-a8d1-449a-8e33-37052dbb1627","Type":"ContainerStarted","Data":"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f"} Oct 03 13:54:35 crc kubenswrapper[4959]: I1003 13:54:35.142030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"774fd066-a8d1-449a-8e33-37052dbb1627","Type":"ContainerStarted","Data":"062a5da45e7aee48cc0feaa8f34c452c68d4bd35bcb03eab23fbb3077b5620f6"} Oct 03 13:54:35 crc kubenswrapper[4959]: I1003 13:54:35.143420 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:35 crc kubenswrapper[4959]: I1003 13:54:35.170765 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.170727969 podStartE2EDuration="2.170727969s" podCreationTimestamp="2025-10-03 13:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:35.167019639 +0000 UTC m=+1444.370363066" watchObservedRunningTime="2025-10-03 13:54:35.170727969 +0000 UTC m=+1444.374071386" Oct 03 13:54:35 crc kubenswrapper[4959]: I1003 13:54:35.694516 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" path="/var/lib/kubelet/pods/fd29d59a-9ff4-4300-b193-84e5b1a72274/volumes" Oct 03 13:54:43 crc kubenswrapper[4959]: I1003 13:54:43.655346 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.137594 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-wr77d"] Oct 03 13:54:44 crc kubenswrapper[4959]: E1003 13:54:44.138118 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="registry-server" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.138141 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="registry-server" Oct 03 13:54:44 crc kubenswrapper[4959]: E1003 13:54:44.138155 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="extract-utilities" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.138161 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="extract-utilities" Oct 03 13:54:44 crc kubenswrapper[4959]: E1003 13:54:44.138181 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="extract-content" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.138208 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="extract-content" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.138432 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd29d59a-9ff4-4300-b193-84e5b1a72274" containerName="registry-server" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.139240 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.141696 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.141792 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.154557 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wr77d"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.306353 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t78z\" (UniqueName: \"kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.306690 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.306721 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.306761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.409635 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.409739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.409807 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.409903 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t78z\" (UniqueName: \"kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.425252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.425298 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.427048 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.431273 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.432652 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: W1003 13:54:44.435698 4959 reflector.go:561] object-"openstack"/"nova-api-config-data": failed to list *v1.Secret: secrets "nova-api-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 03 13:54:44 crc kubenswrapper[4959]: E1003 13:54:44.435759 4959 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"nova-api-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nova-api-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.435800 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t78z\" (UniqueName: \"kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z\") pod \"nova-cell0-cell-mapping-wr77d\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.485184 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.506937 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.507015 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.507093 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.511327 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.512594 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.514903 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.521742 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.537884 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.559750 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613503 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613575 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k99c7\" (UniqueName: \"kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613613 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xnj6\" (UniqueName: \"kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613631 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613650 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613718 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613747 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613768 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxk7w\" (UniqueName: \"kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613796 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.613817 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.618267 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.619440 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.624417 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.680129 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.718264 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.718678 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.719533 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720350 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k99c7\" (UniqueName: \"kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx78n\" (UniqueName: \"kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xnj6\" (UniqueName: \"kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720458 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720538 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720570 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720591 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720657 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxk7w\" (UniqueName: \"kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720681 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.720706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.721568 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.734716 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.761980 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.762265 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.763922 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.764839 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.766154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.770713 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.798769 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k99c7\" (UniqueName: \"kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7\") pod \"nova-metadata-0\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.799867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xnj6\" (UniqueName: \"kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.800453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxk7w\" (UniqueName: \"kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w\") pod \"nova-scheduler-0\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.814721 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.825845 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx78n\" (UniqueName: \"kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.825940 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.826022 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.830362 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.830734 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.842359 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx78n\" (UniqueName: \"kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n\") pod \"nova-cell1-novncproxy-0\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.927861 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.928550 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.928761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.928892 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zd2p\" (UniqueName: \"kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.929008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.929152 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.971992 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.985992 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:54:44 crc kubenswrapper[4959]: I1003 13:54:44.989321 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.031357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.031713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.031868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zd2p\" (UniqueName: \"kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.031990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.032142 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.032316 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.033539 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.034284 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.034907 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.035937 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.036636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.056684 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zd2p\" (UniqueName: \"kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p\") pod \"dnsmasq-dns-845d6d6f59-hstfl\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.103063 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.127624 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-wr77d"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.248262 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wr77d" event={"ID":"20ec93e9-4e68-4d03-9ded-91a642ea9156","Type":"ContainerStarted","Data":"d5895e9ebdeaba35d8cebf4534a7ede13ed6a72298748dc7ce6dfb7d5f4dd9aa"} Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.392787 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lpc2k"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.393995 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.397148 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.401277 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.419022 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lpc2k"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.482117 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.482202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkx4\" (UniqueName: \"kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.482331 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.482398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.537374 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.547154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data\") pod \"nova-api-0\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " pod="openstack/nova-api-0" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.562288 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.579618 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.584076 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.584311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.584409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.584472 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkx4\" (UniqueName: \"kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.587870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.588365 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.588564 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.602127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkx4\" (UniqueName: \"kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4\") pod \"nova-cell1-conductor-db-sync-lpc2k\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: W1003 13:54:45.701330 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb526ecc0_1b95_46d9_b02e_b1e688f5c969.slice/crio-b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a WatchSource:0}: Error finding container b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a: Status 404 returned error can't find the container with id b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.703701 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.712581 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.713019 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:45 crc kubenswrapper[4959]: I1003 13:54:45.736842 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.094671 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:46 crc kubenswrapper[4959]: W1003 13:54:46.101349 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c792940_77cf_4c13_ba14_6d99d0b1281f.slice/crio-96ce7d0d0f9405370e99315fb88853b282cd0ea4b8dbee695f733204d828e55d WatchSource:0}: Error finding container 96ce7d0d0f9405370e99315fb88853b282cd0ea4b8dbee695f733204d828e55d: Status 404 returned error can't find the container with id 96ce7d0d0f9405370e99315fb88853b282cd0ea4b8dbee695f733204d828e55d Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.205749 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lpc2k"] Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.267387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerStarted","Data":"dc4289968451daa1c0d66cb7db694f5a475d005820c4fe4157988248bc3cf3c4"} Oct 03 13:54:46 crc kubenswrapper[4959]: W1003 13:54:46.271131 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5960139e_48e3_4a67_981f_539d78fb46a1.slice/crio-da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c WatchSource:0}: Error finding container da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c: Status 404 returned error can't find the container with id da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.273401 4959 generic.go:334] "Generic (PLEG): container finished" podID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerID="6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4" exitCode=0 Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.273468 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" event={"ID":"5f6e101d-a3cf-4573-8c49-a8b81b9f0766","Type":"ContainerDied","Data":"6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.274796 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" event={"ID":"5f6e101d-a3cf-4573-8c49-a8b81b9f0766","Type":"ContainerStarted","Data":"f5b239987c348b417302d77a22d43946766f904e48ba70d16d9cb0d5fe4156fb"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.279639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f6c1b79-3645-4721-a7cf-6874da9991cd","Type":"ContainerStarted","Data":"da57b4faa30e6cce63f194892e593871777ace69afc126f777d6a4a5698d0be6"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.283487 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerStarted","Data":"96ce7d0d0f9405370e99315fb88853b282cd0ea4b8dbee695f733204d828e55d"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.294920 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b526ecc0-1b95-46d9-b02e-b1e688f5c969","Type":"ContainerStarted","Data":"b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.298239 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wr77d" event={"ID":"20ec93e9-4e68-4d03-9ded-91a642ea9156","Type":"ContainerStarted","Data":"0aca4e4e5b11a2f5ad93029ec98a5edd017bc0bff5d499883f3af4332959c29e"} Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.320085 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-wr77d" podStartSLOduration=2.320068989 podStartE2EDuration="2.320068989s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:46.316635065 +0000 UTC m=+1455.519978482" watchObservedRunningTime="2025-10-03 13:54:46.320068989 +0000 UTC m=+1455.523412406" Oct 03 13:54:46 crc kubenswrapper[4959]: I1003 13:54:46.586695 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.309316 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" event={"ID":"5f6e101d-a3cf-4573-8c49-a8b81b9f0766","Type":"ContainerStarted","Data":"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f"} Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.309637 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.313283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" event={"ID":"5960139e-48e3-4a67-981f-539d78fb46a1","Type":"ContainerStarted","Data":"b54e17dcdfb09d6aeb29e71dcf080822e14a3bd78cc63a8d522dcef2f390bb8d"} Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.313324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" event={"ID":"5960139e-48e3-4a67-981f-539d78fb46a1","Type":"ContainerStarted","Data":"da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c"} Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.343835 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" podStartSLOduration=3.343815782 podStartE2EDuration="3.343815782s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:47.336376641 +0000 UTC m=+1456.539720058" watchObservedRunningTime="2025-10-03 13:54:47.343815782 +0000 UTC m=+1456.547159189" Oct 03 13:54:47 crc kubenswrapper[4959]: I1003 13:54:47.360970 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" podStartSLOduration=2.360947297 podStartE2EDuration="2.360947297s" podCreationTimestamp="2025-10-03 13:54:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:47.350262048 +0000 UTC m=+1456.553605465" watchObservedRunningTime="2025-10-03 13:54:47.360947297 +0000 UTC m=+1456.564290714" Oct 03 13:54:48 crc kubenswrapper[4959]: I1003 13:54:48.322811 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:54:48 crc kubenswrapper[4959]: I1003 13:54:48.332019 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.333235 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f6c1b79-3645-4721-a7cf-6874da9991cd","Type":"ContainerStarted","Data":"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.336480 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerStarted","Data":"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.336527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerStarted","Data":"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.339010 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b526ecc0-1b95-46d9-b02e-b1e688f5c969","Type":"ContainerStarted","Data":"286c3a4d76e172ce7cfa4c2cf6b0f1d71271dbefac5d465aa78b0eb9555090f9"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.339085 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://286c3a4d76e172ce7cfa4c2cf6b0f1d71271dbefac5d465aa78b0eb9555090f9" gracePeriod=30 Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.341860 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerStarted","Data":"f9a14d3f342cd387cfde7a61857031e7aba2a3acb904f84b2b45d694964778f3"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.341902 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerStarted","Data":"35e753ac6de1e60008c5fef25525ceb59336b6d2eaf14a538f17a9e256f1f75b"} Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.341999 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-log" containerID="cri-o://35e753ac6de1e60008c5fef25525ceb59336b6d2eaf14a538f17a9e256f1f75b" gracePeriod=30 Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.342011 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-metadata" containerID="cri-o://f9a14d3f342cd387cfde7a61857031e7aba2a3acb904f84b2b45d694964778f3" gracePeriod=30 Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.350440 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.353483499 podStartE2EDuration="5.350419406s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="2025-10-03 13:54:45.567592508 +0000 UTC m=+1454.770935925" lastFinishedPulling="2025-10-03 13:54:48.564528415 +0000 UTC m=+1457.767871832" observedRunningTime="2025-10-03 13:54:49.349824682 +0000 UTC m=+1458.553168119" watchObservedRunningTime="2025-10-03 13:54:49.350419406 +0000 UTC m=+1458.553762823" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.384781 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.925333286 podStartE2EDuration="5.384764729s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="2025-10-03 13:54:46.104152719 +0000 UTC m=+1455.307496136" lastFinishedPulling="2025-10-03 13:54:48.563584162 +0000 UTC m=+1457.766927579" observedRunningTime="2025-10-03 13:54:49.378151589 +0000 UTC m=+1458.581495016" watchObservedRunningTime="2025-10-03 13:54:49.384764729 +0000 UTC m=+1458.588108146" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.407530 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.551799672 podStartE2EDuration="5.407513062s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="2025-10-03 13:54:45.709265176 +0000 UTC m=+1454.912608593" lastFinishedPulling="2025-10-03 13:54:48.564978566 +0000 UTC m=+1457.768321983" observedRunningTime="2025-10-03 13:54:49.397120379 +0000 UTC m=+1458.600463796" watchObservedRunningTime="2025-10-03 13:54:49.407513062 +0000 UTC m=+1458.610856479" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.427393 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.385788623 podStartE2EDuration="5.427376014s" podCreationTimestamp="2025-10-03 13:54:44 +0000 UTC" firstStartedPulling="2025-10-03 13:54:45.567849024 +0000 UTC m=+1454.771192441" lastFinishedPulling="2025-10-03 13:54:48.609436385 +0000 UTC m=+1457.812779832" observedRunningTime="2025-10-03 13:54:49.416988001 +0000 UTC m=+1458.620331428" watchObservedRunningTime="2025-10-03 13:54:49.427376014 +0000 UTC m=+1458.630719431" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.949310 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.951239 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.961866 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.973559 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.983291 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.983346 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfqjl\" (UniqueName: \"kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.983640 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.987098 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.987143 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:54:49 crc kubenswrapper[4959]: I1003 13:54:49.989817 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.085248 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.085788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.085818 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.085969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfqjl\" (UniqueName: \"kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.086329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.110619 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfqjl\" (UniqueName: \"kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl\") pod \"redhat-marketplace-jk8hq\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.293456 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.377727 4959 generic.go:334] "Generic (PLEG): container finished" podID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerID="35e753ac6de1e60008c5fef25525ceb59336b6d2eaf14a538f17a9e256f1f75b" exitCode=143 Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.378156 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerDied","Data":"35e753ac6de1e60008c5fef25525ceb59336b6d2eaf14a538f17a9e256f1f75b"} Oct 03 13:54:50 crc kubenswrapper[4959]: W1003 13:54:50.818510 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e7c8471_1c50_4862_83e6_eb27f7cdb788.slice/crio-0a29c737d29ba08e00f8bf61ca6da3faaa63f368f24b407115e8cfd1c81a7454 WatchSource:0}: Error finding container 0a29c737d29ba08e00f8bf61ca6da3faaa63f368f24b407115e8cfd1c81a7454: Status 404 returned error can't find the container with id 0a29c737d29ba08e00f8bf61ca6da3faaa63f368f24b407115e8cfd1c81a7454 Oct 03 13:54:50 crc kubenswrapper[4959]: I1003 13:54:50.818655 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.320057 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.320342 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" containerName="kube-state-metrics" containerID="cri-o://8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345" gracePeriod=30 Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.388104 4959 generic.go:334] "Generic (PLEG): container finished" podID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerID="3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16" exitCode=0 Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.388163 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerDied","Data":"3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16"} Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.388410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerStarted","Data":"0a29c737d29ba08e00f8bf61ca6da3faaa63f368f24b407115e8cfd1c81a7454"} Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.781181 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.815765 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6xtw\" (UniqueName: \"kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw\") pod \"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a\" (UID: \"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a\") " Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.825400 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw" (OuterVolumeSpecName: "kube-api-access-k6xtw") pod "1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" (UID: "1bc8649d-27d5-4834-bc5d-d46f32cbdd0a"). InnerVolumeSpecName "kube-api-access-k6xtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:51 crc kubenswrapper[4959]: I1003 13:54:51.917961 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6xtw\" (UniqueName: \"kubernetes.io/projected/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a-kube-api-access-k6xtw\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.399170 4959 generic.go:334] "Generic (PLEG): container finished" podID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" containerID="8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345" exitCode=2 Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.399221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a","Type":"ContainerDied","Data":"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345"} Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.399249 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.399264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bc8649d-27d5-4834-bc5d-d46f32cbdd0a","Type":"ContainerDied","Data":"fc03974bba56fdd16f826294c190b9492cd471046c9044639a23747af62a869c"} Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.399287 4959 scope.go:117] "RemoveContainer" containerID="8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.401945 4959 generic.go:334] "Generic (PLEG): container finished" podID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerID="529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b" exitCode=0 Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.401988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerDied","Data":"529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b"} Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.427413 4959 scope.go:117] "RemoveContainer" containerID="8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345" Oct 03 13:54:52 crc kubenswrapper[4959]: E1003 13:54:52.429276 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345\": container with ID starting with 8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345 not found: ID does not exist" containerID="8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.429325 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345"} err="failed to get container status \"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345\": rpc error: code = NotFound desc = could not find container \"8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345\": container with ID starting with 8a0b84dff2ee849f2be2be5ab20667bae6bbce2f838427b461f848c7a5020345 not found: ID does not exist" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.451670 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.464237 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.483101 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:52 crc kubenswrapper[4959]: E1003 13:54:52.483610 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" containerName="kube-state-metrics" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.483629 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" containerName="kube-state-metrics" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.483844 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" containerName="kube-state-metrics" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.484594 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.486607 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.486740 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.488436 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.529722 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.529791 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqtlg\" (UniqueName: \"kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.529812 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.529834 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.631718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.631801 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqtlg\" (UniqueName: \"kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.631831 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.631863 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.637232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.637936 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.641679 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.655055 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqtlg\" (UniqueName: \"kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg\") pod \"kube-state-metrics-0\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " pod="openstack/kube-state-metrics-0" Oct 03 13:54:52 crc kubenswrapper[4959]: I1003 13:54:52.803926 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.293921 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.297017 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-central-agent" containerID="cri-o://3b2ab7b13434044306549d54bc915d4778e56be24c7855da96228c6a69b18300" gracePeriod=30 Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.297226 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="proxy-httpd" containerID="cri-o://5cd4b2ad25a3ac5039c4b68acd1791fc1fc99a9b43b72b53b276f124d7531433" gracePeriod=30 Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.297269 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="sg-core" containerID="cri-o://5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c" gracePeriod=30 Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.297299 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-notification-agent" containerID="cri-o://5bb3fa231f29ee6dab17213f0ac78b3f8a0b41725add6f76cf68a89d2897b04e" gracePeriod=30 Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.314350 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.415830 4959 generic.go:334] "Generic (PLEG): container finished" podID="20ec93e9-4e68-4d03-9ded-91a642ea9156" containerID="0aca4e4e5b11a2f5ad93029ec98a5edd017bc0bff5d499883f3af4332959c29e" exitCode=0 Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.415922 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wr77d" event={"ID":"20ec93e9-4e68-4d03-9ded-91a642ea9156","Type":"ContainerDied","Data":"0aca4e4e5b11a2f5ad93029ec98a5edd017bc0bff5d499883f3af4332959c29e"} Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.424523 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"593264f5-6188-4861-9d83-6918034eb46e","Type":"ContainerStarted","Data":"f4bde2bb7cc603e46e0fb2ffd5621aa392edc67c8c3939960711728173c6543b"} Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.431089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerStarted","Data":"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a"} Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.463126 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jk8hq" podStartSLOduration=2.908053711 podStartE2EDuration="4.463109658s" podCreationTimestamp="2025-10-03 13:54:49 +0000 UTC" firstStartedPulling="2025-10-03 13:54:51.390625885 +0000 UTC m=+1460.593969322" lastFinishedPulling="2025-10-03 13:54:52.945681832 +0000 UTC m=+1462.149025269" observedRunningTime="2025-10-03 13:54:53.460082145 +0000 UTC m=+1462.663425582" watchObservedRunningTime="2025-10-03 13:54:53.463109658 +0000 UTC m=+1462.666453075" Oct 03 13:54:53 crc kubenswrapper[4959]: E1003 13:54:53.494157 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82865657_4141_487e_8469_546f8b334bcd.slice/crio-5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ec93e9_4e68_4d03_9ded_91a642ea9156.slice/crio-conmon-0aca4e4e5b11a2f5ad93029ec98a5edd017bc0bff5d499883f3af4332959c29e.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:54:53 crc kubenswrapper[4959]: I1003 13:54:53.697499 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc8649d-27d5-4834-bc5d-d46f32cbdd0a" path="/var/lib/kubelet/pods/1bc8649d-27d5-4834-bc5d-d46f32cbdd0a/volumes" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443001 4959 generic.go:334] "Generic (PLEG): container finished" podID="82865657-4141-487e-8469-546f8b334bcd" containerID="5cd4b2ad25a3ac5039c4b68acd1791fc1fc99a9b43b72b53b276f124d7531433" exitCode=0 Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443564 4959 generic.go:334] "Generic (PLEG): container finished" podID="82865657-4141-487e-8469-546f8b334bcd" containerID="5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c" exitCode=2 Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443574 4959 generic.go:334] "Generic (PLEG): container finished" podID="82865657-4141-487e-8469-546f8b334bcd" containerID="5bb3fa231f29ee6dab17213f0ac78b3f8a0b41725add6f76cf68a89d2897b04e" exitCode=0 Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443585 4959 generic.go:334] "Generic (PLEG): container finished" podID="82865657-4141-487e-8469-546f8b334bcd" containerID="3b2ab7b13434044306549d54bc915d4778e56be24c7855da96228c6a69b18300" exitCode=0 Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443104 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerDied","Data":"5cd4b2ad25a3ac5039c4b68acd1791fc1fc99a9b43b72b53b276f124d7531433"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443650 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerDied","Data":"5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerDied","Data":"5bb3fa231f29ee6dab17213f0ac78b3f8a0b41725add6f76cf68a89d2897b04e"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.443674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerDied","Data":"3b2ab7b13434044306549d54bc915d4778e56be24c7855da96228c6a69b18300"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.445339 4959 generic.go:334] "Generic (PLEG): container finished" podID="5960139e-48e3-4a67-981f-539d78fb46a1" containerID="b54e17dcdfb09d6aeb29e71dcf080822e14a3bd78cc63a8d522dcef2f390bb8d" exitCode=0 Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.445411 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" event={"ID":"5960139e-48e3-4a67-981f-539d78fb46a1","Type":"ContainerDied","Data":"b54e17dcdfb09d6aeb29e71dcf080822e14a3bd78cc63a8d522dcef2f390bb8d"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.451273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"593264f5-6188-4861-9d83-6918034eb46e","Type":"ContainerStarted","Data":"437ecf2e1e223998fb6b040d3dc342771553ad152a649a307ba005f342deed15"} Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.451347 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.487713 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.06604003 podStartE2EDuration="2.487693092s" podCreationTimestamp="2025-10-03 13:54:52 +0000 UTC" firstStartedPulling="2025-10-03 13:54:53.335345828 +0000 UTC m=+1462.538689245" lastFinishedPulling="2025-10-03 13:54:53.75699889 +0000 UTC m=+1462.960342307" observedRunningTime="2025-10-03 13:54:54.475484945 +0000 UTC m=+1463.678828372" watchObservedRunningTime="2025-10-03 13:54:54.487693092 +0000 UTC m=+1463.691036509" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.812642 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.828563 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.877991 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.878483 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.878536 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts\") pod \"20ec93e9-4e68-4d03-9ded-91a642ea9156\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.878613 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x5wg\" (UniqueName: \"kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879256 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879278 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879399 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879424 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879449 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle\") pod \"20ec93e9-4e68-4d03-9ded-91a642ea9156\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879473 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t78z\" (UniqueName: \"kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z\") pod \"20ec93e9-4e68-4d03-9ded-91a642ea9156\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879566 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd\") pod \"82865657-4141-487e-8469-546f8b334bcd\" (UID: \"82865657-4141-487e-8469-546f8b334bcd\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.879623 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data\") pod \"20ec93e9-4e68-4d03-9ded-91a642ea9156\" (UID: \"20ec93e9-4e68-4d03-9ded-91a642ea9156\") " Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.880055 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.881454 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.885529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z" (OuterVolumeSpecName: "kube-api-access-7t78z") pod "20ec93e9-4e68-4d03-9ded-91a642ea9156" (UID: "20ec93e9-4e68-4d03-9ded-91a642ea9156"). InnerVolumeSpecName "kube-api-access-7t78z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.900434 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg" (OuterVolumeSpecName: "kube-api-access-6x5wg") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "kube-api-access-6x5wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.903607 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts" (OuterVolumeSpecName: "scripts") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.911453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts" (OuterVolumeSpecName: "scripts") pod "20ec93e9-4e68-4d03-9ded-91a642ea9156" (UID: "20ec93e9-4e68-4d03-9ded-91a642ea9156"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.933848 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data" (OuterVolumeSpecName: "config-data") pod "20ec93e9-4e68-4d03-9ded-91a642ea9156" (UID: "20ec93e9-4e68-4d03-9ded-91a642ea9156"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.939438 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.943505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20ec93e9-4e68-4d03-9ded-91a642ea9156" (UID: "20ec93e9-4e68-4d03-9ded-91a642ea9156"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.973794 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.981942 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82865657-4141-487e-8469-546f8b334bcd-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.981977 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982219 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982234 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x5wg\" (UniqueName: \"kubernetes.io/projected/82865657-4141-487e-8469-546f8b334bcd-kube-api-access-6x5wg\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982247 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982257 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982268 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ec93e9-4e68-4d03-9ded-91a642ea9156-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.982308 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t78z\" (UniqueName: \"kubernetes.io/projected/20ec93e9-4e68-4d03-9ded-91a642ea9156-kube-api-access-7t78z\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:54 crc kubenswrapper[4959]: I1003 13:54:54.994777 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.008076 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.030405 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data" (OuterVolumeSpecName: "config-data") pod "82865657-4141-487e-8469-546f8b334bcd" (UID: "82865657-4141-487e-8469-546f8b334bcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.084573 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.084607 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82865657-4141-487e-8469-546f8b334bcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.105319 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.177073 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.462093 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-wr77d" event={"ID":"20ec93e9-4e68-4d03-9ded-91a642ea9156","Type":"ContainerDied","Data":"d5895e9ebdeaba35d8cebf4534a7ede13ed6a72298748dc7ce6dfb7d5f4dd9aa"} Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.462130 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5895e9ebdeaba35d8cebf4534a7ede13ed6a72298748dc7ce6dfb7d5f4dd9aa" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.462291 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-wr77d" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.480444 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82865657-4141-487e-8469-546f8b334bcd","Type":"ContainerDied","Data":"22d9ec904cb11528b29cfa20d9b5ba6a378bb62076311a3d7e3fa18bd4947741"} Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.480523 4959 scope.go:117] "RemoveContainer" containerID="5cd4b2ad25a3ac5039c4b68acd1791fc1fc99a9b43b72b53b276f124d7531433" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.480585 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="dnsmasq-dns" containerID="cri-o://de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942" gracePeriod=10 Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.480945 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.511858 4959 scope.go:117] "RemoveContainer" containerID="5ae58f06856209a7293c5e5560811e8da03188ee2cfad58b85f603d7b665544c" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.524562 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.534146 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.535639 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546044 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:55 crc kubenswrapper[4959]: E1003 13:54:55.546502 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-central-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546519 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-central-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: E1003 13:54:55.546531 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-notification-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546539 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-notification-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: E1003 13:54:55.546549 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="sg-core" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546557 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="sg-core" Oct 03 13:54:55 crc kubenswrapper[4959]: E1003 13:54:55.546592 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ec93e9-4e68-4d03-9ded-91a642ea9156" containerName="nova-manage" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546600 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ec93e9-4e68-4d03-9ded-91a642ea9156" containerName="nova-manage" Oct 03 13:54:55 crc kubenswrapper[4959]: E1003 13:54:55.546614 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="proxy-httpd" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546622 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="proxy-httpd" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546792 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-notification-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546805 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="sg-core" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546820 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="ceilometer-central-agent" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546834 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ec93e9-4e68-4d03-9ded-91a642ea9156" containerName="nova-manage" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.546847 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82865657-4141-487e-8469-546f8b334bcd" containerName="proxy-httpd" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.548517 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.550612 4959 scope.go:117] "RemoveContainer" containerID="5bb3fa231f29ee6dab17213f0ac78b3f8a0b41725add6f76cf68a89d2897b04e" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.551131 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.551238 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.552045 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.554416 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.610538 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.610596 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.610747 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.610825 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.610858 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz5w8\" (UniqueName: \"kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.611731 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.611807 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.611889 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.655126 4959 scope.go:117] "RemoveContainer" containerID="3b2ab7b13434044306549d54bc915d4778e56be24c7855da96228c6a69b18300" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.706676 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82865657-4141-487e-8469-546f8b334bcd" path="/var/lib/kubelet/pods/82865657-4141-487e-8469-546f8b334bcd/volumes" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.707690 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.707959 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-log" containerID="cri-o://22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" gracePeriod=30 Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.708023 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-api" containerID="cri-o://fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" gracePeriod=30 Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.713741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.713856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.713936 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz5w8\" (UniqueName: \"kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.715500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.715605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.715705 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.715751 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.715775 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.718838 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.722931 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.723417 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.729810 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.730251 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.730476 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.731829 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.747293 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz5w8\" (UniqueName: \"kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8\") pod \"ceilometer-0\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " pod="openstack/ceilometer-0" Oct 03 13:54:55 crc kubenswrapper[4959]: I1003 13:54:55.957350 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.027344 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.168940 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.174683 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224015 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224081 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data\") pod \"5960139e-48e3-4a67-981f-539d78fb46a1\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224172 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts\") pod \"5960139e-48e3-4a67-981f-539d78fb46a1\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224226 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224263 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkjqc\" (UniqueName: \"kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224299 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224352 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.224380 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfkx4\" (UniqueName: \"kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4\") pod \"5960139e-48e3-4a67-981f-539d78fb46a1\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.225532 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle\") pod \"5960139e-48e3-4a67-981f-539d78fb46a1\" (UID: \"5960139e-48e3-4a67-981f-539d78fb46a1\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.225612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc\") pod \"b13a531c-8bcc-46b4-b034-525a5774081d\" (UID: \"b13a531c-8bcc-46b4-b034-525a5774081d\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.232715 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc" (OuterVolumeSpecName: "kube-api-access-pkjqc") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "kube-api-access-pkjqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.249488 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4" (OuterVolumeSpecName: "kube-api-access-gfkx4") pod "5960139e-48e3-4a67-981f-539d78fb46a1" (UID: "5960139e-48e3-4a67-981f-539d78fb46a1"). InnerVolumeSpecName "kube-api-access-gfkx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.250926 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts" (OuterVolumeSpecName: "scripts") pod "5960139e-48e3-4a67-981f-539d78fb46a1" (UID: "5960139e-48e3-4a67-981f-539d78fb46a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.290937 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.308791 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.329804 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xnj6\" (UniqueName: \"kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6\") pod \"8c792940-77cf-4c13-ba14-6d99d0b1281f\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.329944 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data\") pod \"8c792940-77cf-4c13-ba14-6d99d0b1281f\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330022 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle\") pod \"8c792940-77cf-4c13-ba14-6d99d0b1281f\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330068 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs\") pod \"8c792940-77cf-4c13-ba14-6d99d0b1281f\" (UID: \"8c792940-77cf-4c13-ba14-6d99d0b1281f\") " Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330574 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330592 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkjqc\" (UniqueName: \"kubernetes.io/projected/b13a531c-8bcc-46b4-b034-525a5774081d-kube-api-access-pkjqc\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330605 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.330616 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfkx4\" (UniqueName: \"kubernetes.io/projected/5960139e-48e3-4a67-981f-539d78fb46a1-kube-api-access-gfkx4\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.332523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs" (OuterVolumeSpecName: "logs") pod "8c792940-77cf-4c13-ba14-6d99d0b1281f" (UID: "8c792940-77cf-4c13-ba14-6d99d0b1281f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.338694 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.344233 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.345031 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data" (OuterVolumeSpecName: "config-data") pod "5960139e-48e3-4a67-981f-539d78fb46a1" (UID: "5960139e-48e3-4a67-981f-539d78fb46a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.345561 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6" (OuterVolumeSpecName: "kube-api-access-7xnj6") pod "8c792940-77cf-4c13-ba14-6d99d0b1281f" (UID: "8c792940-77cf-4c13-ba14-6d99d0b1281f"). InnerVolumeSpecName "kube-api-access-7xnj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.354480 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.364293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5960139e-48e3-4a67-981f-539d78fb46a1" (UID: "5960139e-48e3-4a67-981f-539d78fb46a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.376832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data" (OuterVolumeSpecName: "config-data") pod "8c792940-77cf-4c13-ba14-6d99d0b1281f" (UID: "8c792940-77cf-4c13-ba14-6d99d0b1281f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.378183 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c792940-77cf-4c13-ba14-6d99d0b1281f" (UID: "8c792940-77cf-4c13-ba14-6d99d0b1281f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.382514 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config" (OuterVolumeSpecName: "config") pod "b13a531c-8bcc-46b4-b034-525a5774081d" (UID: "b13a531c-8bcc-46b4-b034-525a5774081d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432539 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432577 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432589 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432598 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c792940-77cf-4c13-ba14-6d99d0b1281f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432608 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c792940-77cf-4c13-ba14-6d99d0b1281f-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432617 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432626 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432634 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5960139e-48e3-4a67-981f-539d78fb46a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432642 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xnj6\" (UniqueName: \"kubernetes.io/projected/8c792940-77cf-4c13-ba14-6d99d0b1281f-kube-api-access-7xnj6\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.432652 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b13a531c-8bcc-46b4-b034-525a5774081d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.491017 4959 generic.go:334] "Generic (PLEG): container finished" podID="b13a531c-8bcc-46b4-b034-525a5774081d" containerID="de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942" exitCode=0 Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.491072 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.491088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" event={"ID":"b13a531c-8bcc-46b4-b034-525a5774081d","Type":"ContainerDied","Data":"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.491114 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-nqkwd" event={"ID":"b13a531c-8bcc-46b4-b034-525a5774081d","Type":"ContainerDied","Data":"9331d28cadff0d1e1ad67e4b6d795ccac93a13113a776c8dc43807710ceff864"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.491131 4959 scope.go:117] "RemoveContainer" containerID="de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.496017 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" event={"ID":"5960139e-48e3-4a67-981f-539d78fb46a1","Type":"ContainerDied","Data":"da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.496056 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da391ffaefbe4df866aee44897742d31740352dcf3d736352f7ccbfe40fa630c" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.496055 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lpc2k" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.510758 4959 generic.go:334] "Generic (PLEG): container finished" podID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerID="fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" exitCode=0 Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.510785 4959 generic.go:334] "Generic (PLEG): container finished" podID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerID="22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" exitCode=143 Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.510829 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.511023 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerDied","Data":"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.511111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerDied","Data":"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.511175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c792940-77cf-4c13-ba14-6d99d0b1281f","Type":"ContainerDied","Data":"96ce7d0d0f9405370e99315fb88853b282cd0ea4b8dbee695f733204d828e55d"} Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.514310 4959 scope.go:117] "RemoveContainer" containerID="aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.528413 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.539509 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-nqkwd"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.552218 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562047 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.562487 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-log" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562506 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-log" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.562523 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5960139e-48e3-4a67-981f-539d78fb46a1" containerName="nova-cell1-conductor-db-sync" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562530 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5960139e-48e3-4a67-981f-539d78fb46a1" containerName="nova-cell1-conductor-db-sync" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.562560 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-api" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562566 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-api" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.562578 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="dnsmasq-dns" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562584 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="dnsmasq-dns" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.562597 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="init" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562602 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="init" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562770 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-api" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562787 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5960139e-48e3-4a67-981f-539d78fb46a1" containerName="nova-cell1-conductor-db-sync" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562801 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" containerName="nova-api-log" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.562810 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" containerName="dnsmasq-dns" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.563544 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.565957 4959 scope.go:117] "RemoveContainer" containerID="de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.566305 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.570705 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942\": container with ID starting with de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942 not found: ID does not exist" containerID="de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.570750 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942"} err="failed to get container status \"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942\": rpc error: code = NotFound desc = could not find container \"de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942\": container with ID starting with de729cb0f9e59c6a8e2f63b81fddba53c02f93f768e6a47d5124eb0127de0942 not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.570772 4959 scope.go:117] "RemoveContainer" containerID="aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.570882 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.571530 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83\": container with ID starting with aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83 not found: ID does not exist" containerID="aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.571563 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83"} err="failed to get container status \"aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83\": rpc error: code = NotFound desc = could not find container \"aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83\": container with ID starting with aa1313ee2c59c9fde00ce5da8d983197a2e4ad80bb4d3543cd758702e4ec9c83 not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.571582 4959 scope.go:117] "RemoveContainer" containerID="fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.579318 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.593479 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.609769 4959 scope.go:117] "RemoveContainer" containerID="22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.635720 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.635780 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lkzg\" (UniqueName: \"kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.635861 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.638479 4959 scope.go:117] "RemoveContainer" containerID="fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.638849 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801\": container with ID starting with fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801 not found: ID does not exist" containerID="fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.638887 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801"} err="failed to get container status \"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801\": rpc error: code = NotFound desc = could not find container \"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801\": container with ID starting with fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801 not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.638915 4959 scope.go:117] "RemoveContainer" containerID="22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" Oct 03 13:54:56 crc kubenswrapper[4959]: E1003 13:54:56.639092 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb\": container with ID starting with 22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb not found: ID does not exist" containerID="22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.639119 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb"} err="failed to get container status \"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb\": rpc error: code = NotFound desc = could not find container \"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb\": container with ID starting with 22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.639134 4959 scope.go:117] "RemoveContainer" containerID="fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.639389 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801"} err="failed to get container status \"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801\": rpc error: code = NotFound desc = could not find container \"fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801\": container with ID starting with fd43e2e0e325db72088b4a8a8333b2f26abc3ae9702be869e5f55e5c8ebc4801 not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.639418 4959 scope.go:117] "RemoveContainer" containerID="22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.639619 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb"} err="failed to get container status \"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb\": rpc error: code = NotFound desc = could not find container \"22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb\": container with ID starting with 22f2430df5cb94a5a80295e43cf1e53e2e9e2db71ce34bba0b2942f7c89fa3cb not found: ID does not exist" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.648122 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.650794 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.654020 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.669517 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.737821 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.738223 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.738348 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.738449 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lkzg\" (UniqueName: \"kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.738592 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.738741 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9n9m\" (UniqueName: \"kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.739380 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.741045 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.743960 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.763069 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lkzg\" (UniqueName: \"kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg\") pod \"nova-cell1-conductor-0\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.841081 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9n9m\" (UniqueName: \"kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.841140 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.841249 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.841267 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.841874 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.844538 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.845735 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.858660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9n9m\" (UniqueName: \"kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m\") pod \"nova-api-0\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " pod="openstack/nova-api-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.897175 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:56 crc kubenswrapper[4959]: I1003 13:54:56.980099 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.521037 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerStarted","Data":"85800a49b72e4df0d36e466e500bca16159ca593f2014e92de316822e8cd451e"} Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.524386 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerName="nova-scheduler-scheduler" containerID="cri-o://ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" gracePeriod=30 Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.563749 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:54:57 crc kubenswrapper[4959]: W1003 13:54:57.568430 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0c0088_66f1_4f85_8a34_a1f34867a37a.slice/crio-043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182 WatchSource:0}: Error finding container 043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182: Status 404 returned error can't find the container with id 043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182 Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.572006 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.701123 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c792940-77cf-4c13-ba14-6d99d0b1281f" path="/var/lib/kubelet/pods/8c792940-77cf-4c13-ba14-6d99d0b1281f/volumes" Oct 03 13:54:57 crc kubenswrapper[4959]: I1003 13:54:57.703006 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13a531c-8bcc-46b4-b034-525a5774081d" path="/var/lib/kubelet/pods/b13a531c-8bcc-46b4-b034-525a5774081d/volumes" Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.543091 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerStarted","Data":"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.543441 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerStarted","Data":"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.543456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerStarted","Data":"c636c885b3e171acb990705278895f26cd7bec04f0e2ab18c7d26c45808d3466"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.544607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerStarted","Data":"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.547516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ff0c0088-66f1-4f85-8a34-a1f34867a37a","Type":"ContainerStarted","Data":"e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.547550 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ff0c0088-66f1-4f85-8a34-a1f34867a37a","Type":"ContainerStarted","Data":"043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182"} Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.547852 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 13:54:58 crc kubenswrapper[4959]: I1003 13:54:58.571676 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.571658107 podStartE2EDuration="2.571658107s" podCreationTimestamp="2025-10-03 13:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:58.561541482 +0000 UTC m=+1467.764884899" watchObservedRunningTime="2025-10-03 13:54:58.571658107 +0000 UTC m=+1467.775001524" Oct 03 13:54:59 crc kubenswrapper[4959]: I1003 13:54:59.559088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerStarted","Data":"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666"} Oct 03 13:54:59 crc kubenswrapper[4959]: E1003 13:54:59.975559 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:54:59 crc kubenswrapper[4959]: E1003 13:54:59.977640 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:54:59 crc kubenswrapper[4959]: E1003 13:54:59.979260 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:54:59 crc kubenswrapper[4959]: E1003 13:54:59.979325 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerName="nova-scheduler-scheduler" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.294792 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.294837 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.339716 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.356100 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.356081959 podStartE2EDuration="4.356081959s" podCreationTimestamp="2025-10-03 13:54:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:54:59.586856343 +0000 UTC m=+1468.790199810" watchObservedRunningTime="2025-10-03 13:55:00.356081959 +0000 UTC m=+1469.559425366" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.569108 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerStarted","Data":"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956"} Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.623669 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.674019 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:55:00 crc kubenswrapper[4959]: I1003 13:55:00.994169 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.011839 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxk7w\" (UniqueName: \"kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w\") pod \"2f6c1b79-3645-4721-a7cf-6874da9991cd\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.011934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle\") pod \"2f6c1b79-3645-4721-a7cf-6874da9991cd\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.012082 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data\") pod \"2f6c1b79-3645-4721-a7cf-6874da9991cd\" (UID: \"2f6c1b79-3645-4721-a7cf-6874da9991cd\") " Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.018814 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w" (OuterVolumeSpecName: "kube-api-access-hxk7w") pod "2f6c1b79-3645-4721-a7cf-6874da9991cd" (UID: "2f6c1b79-3645-4721-a7cf-6874da9991cd"). InnerVolumeSpecName "kube-api-access-hxk7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.049390 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f6c1b79-3645-4721-a7cf-6874da9991cd" (UID: "2f6c1b79-3645-4721-a7cf-6874da9991cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.068557 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data" (OuterVolumeSpecName: "config-data") pod "2f6c1b79-3645-4721-a7cf-6874da9991cd" (UID: "2f6c1b79-3645-4721-a7cf-6874da9991cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.114011 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.114074 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxk7w\" (UniqueName: \"kubernetes.io/projected/2f6c1b79-3645-4721-a7cf-6874da9991cd-kube-api-access-hxk7w\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.114088 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6c1b79-3645-4721-a7cf-6874da9991cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.579941 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerStarted","Data":"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af"} Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.580294 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.582503 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" exitCode=0 Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.582566 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.582554 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f6c1b79-3645-4721-a7cf-6874da9991cd","Type":"ContainerDied","Data":"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28"} Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.582744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f6c1b79-3645-4721-a7cf-6874da9991cd","Type":"ContainerDied","Data":"da57b4faa30e6cce63f194892e593871777ace69afc126f777d6a4a5698d0be6"} Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.582774 4959 scope.go:117] "RemoveContainer" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.606446 4959 scope.go:117] "RemoveContainer" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" Oct 03 13:55:01 crc kubenswrapper[4959]: E1003 13:55:01.607106 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28\": container with ID starting with ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28 not found: ID does not exist" containerID="ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.607159 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28"} err="failed to get container status \"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28\": rpc error: code = NotFound desc = could not find container \"ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28\": container with ID starting with ceee279518a7b0fc3c0ebbd81e4c0d399a24212621e052a1a7ecc9148fce5b28 not found: ID does not exist" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.625464 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.989464462 podStartE2EDuration="6.625433853s" podCreationTimestamp="2025-10-03 13:54:55 +0000 UTC" firstStartedPulling="2025-10-03 13:54:56.565484243 +0000 UTC m=+1465.768827660" lastFinishedPulling="2025-10-03 13:55:01.201453634 +0000 UTC m=+1470.404797051" observedRunningTime="2025-10-03 13:55:01.612455318 +0000 UTC m=+1470.815798755" watchObservedRunningTime="2025-10-03 13:55:01.625433853 +0000 UTC m=+1470.828777270" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.638404 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.652925 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.664598 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:01 crc kubenswrapper[4959]: E1003 13:55:01.665220 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerName="nova-scheduler-scheduler" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.665243 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerName="nova-scheduler-scheduler" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.665477 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" containerName="nova-scheduler-scheduler" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.666313 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.669944 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.674818 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.704261 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f6c1b79-3645-4721-a7cf-6874da9991cd" path="/var/lib/kubelet/pods/2f6c1b79-3645-4721-a7cf-6874da9991cd/volumes" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.727398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.727524 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khf66\" (UniqueName: \"kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.727552 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.829394 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khf66\" (UniqueName: \"kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.829449 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.829546 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.833253 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.838344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.845187 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khf66\" (UniqueName: \"kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66\") pod \"nova-scheduler-0\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:01 crc kubenswrapper[4959]: I1003 13:55:01.981991 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:02 crc kubenswrapper[4959]: I1003 13:55:02.409425 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:02 crc kubenswrapper[4959]: W1003 13:55:02.417602 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5659216_fc65_4dbd_a13b_addef38f4ebd.slice/crio-b4aaa741d1e8ac948c3df7d7539cd5c5917034c6f3244fa3d8d1ec7665a9b5ff WatchSource:0}: Error finding container b4aaa741d1e8ac948c3df7d7539cd5c5917034c6f3244fa3d8d1ec7665a9b5ff: Status 404 returned error can't find the container with id b4aaa741d1e8ac948c3df7d7539cd5c5917034c6f3244fa3d8d1ec7665a9b5ff Oct 03 13:55:02 crc kubenswrapper[4959]: I1003 13:55:02.596465 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5659216-fc65-4dbd-a13b-addef38f4ebd","Type":"ContainerStarted","Data":"b4aaa741d1e8ac948c3df7d7539cd5c5917034c6f3244fa3d8d1ec7665a9b5ff"} Oct 03 13:55:02 crc kubenswrapper[4959]: I1003 13:55:02.596892 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jk8hq" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="registry-server" containerID="cri-o://b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a" gracePeriod=2 Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:02.826572 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.053445 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.156565 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content\") pod \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.156705 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities\") pod \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.156865 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfqjl\" (UniqueName: \"kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl\") pod \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\" (UID: \"4e7c8471-1c50-4862-83e6-eb27f7cdb788\") " Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.158102 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities" (OuterVolumeSpecName: "utilities") pod "4e7c8471-1c50-4862-83e6-eb27f7cdb788" (UID: "4e7c8471-1c50-4862-83e6-eb27f7cdb788"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.161156 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl" (OuterVolumeSpecName: "kube-api-access-kfqjl") pod "4e7c8471-1c50-4862-83e6-eb27f7cdb788" (UID: "4e7c8471-1c50-4862-83e6-eb27f7cdb788"). InnerVolumeSpecName "kube-api-access-kfqjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.172830 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e7c8471-1c50-4862-83e6-eb27f7cdb788" (UID: "4e7c8471-1c50-4862-83e6-eb27f7cdb788"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.259394 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfqjl\" (UniqueName: \"kubernetes.io/projected/4e7c8471-1c50-4862-83e6-eb27f7cdb788-kube-api-access-kfqjl\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.259437 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.259447 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e7c8471-1c50-4862-83e6-eb27f7cdb788-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.616910 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5659216-fc65-4dbd-a13b-addef38f4ebd","Type":"ContainerStarted","Data":"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7"} Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.646736 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6467162440000003 podStartE2EDuration="2.646716244s" podCreationTimestamp="2025-10-03 13:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:03.641410875 +0000 UTC m=+1472.844754292" watchObservedRunningTime="2025-10-03 13:55:03.646716244 +0000 UTC m=+1472.850059661" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.647597 4959 generic.go:334] "Generic (PLEG): container finished" podID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerID="b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a" exitCode=0 Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.647629 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerDied","Data":"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a"} Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.647652 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk8hq" event={"ID":"4e7c8471-1c50-4862-83e6-eb27f7cdb788","Type":"ContainerDied","Data":"0a29c737d29ba08e00f8bf61ca6da3faaa63f368f24b407115e8cfd1c81a7454"} Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.647667 4959 scope.go:117] "RemoveContainer" containerID="b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.647824 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk8hq" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.694304 4959 scope.go:117] "RemoveContainer" containerID="529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.707506 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.723785 4959 scope.go:117] "RemoveContainer" containerID="3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.734765 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk8hq"] Oct 03 13:55:03 crc kubenswrapper[4959]: E1003 13:55:03.761397 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e7c8471_1c50_4862_83e6_eb27f7cdb788.slice\": RecentStats: unable to find data in memory cache]" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.792308 4959 scope.go:117] "RemoveContainer" containerID="b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a" Oct 03 13:55:03 crc kubenswrapper[4959]: E1003 13:55:03.793995 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a\": container with ID starting with b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a not found: ID does not exist" containerID="b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.794032 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a"} err="failed to get container status \"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a\": rpc error: code = NotFound desc = could not find container \"b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a\": container with ID starting with b4466ed40f7fcd898524fa0bea0f6f0f66e2c2393c0b247fb0ec22501d94e37a not found: ID does not exist" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.794054 4959 scope.go:117] "RemoveContainer" containerID="529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b" Oct 03 13:55:03 crc kubenswrapper[4959]: E1003 13:55:03.794464 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b\": container with ID starting with 529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b not found: ID does not exist" containerID="529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.794494 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b"} err="failed to get container status \"529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b\": rpc error: code = NotFound desc = could not find container \"529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b\": container with ID starting with 529610a97877997954ce43cc5fcf0577595f3f26dd4f7277b6058ae47482a24b not found: ID does not exist" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.794511 4959 scope.go:117] "RemoveContainer" containerID="3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16" Oct 03 13:55:03 crc kubenswrapper[4959]: E1003 13:55:03.794803 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16\": container with ID starting with 3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16 not found: ID does not exist" containerID="3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16" Oct 03 13:55:03 crc kubenswrapper[4959]: I1003 13:55:03.794831 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16"} err="failed to get container status \"3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16\": rpc error: code = NotFound desc = could not find container \"3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16\": container with ID starting with 3c0fff8aa1c5d8438b8712836a7158d5c4fa287a1a5d31bbce47130c44e83e16 not found: ID does not exist" Oct 03 13:55:05 crc kubenswrapper[4959]: I1003 13:55:05.697534 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" path="/var/lib/kubelet/pods/4e7c8471-1c50-4862-83e6-eb27f7cdb788/volumes" Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.044553 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.044624 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.930668 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.981278 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.981361 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:06 crc kubenswrapper[4959]: I1003 13:55:06.982659 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:55:08 crc kubenswrapper[4959]: I1003 13:55:08.022611 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:08 crc kubenswrapper[4959]: I1003 13:55:08.064389 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:11 crc kubenswrapper[4959]: I1003 13:55:11.983122 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:55:12 crc kubenswrapper[4959]: I1003 13:55:12.015447 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:55:12 crc kubenswrapper[4959]: I1003 13:55:12.782347 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:55:16 crc kubenswrapper[4959]: I1003 13:55:16.984417 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:55:16 crc kubenswrapper[4959]: I1003 13:55:16.985602 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:55:16 crc kubenswrapper[4959]: I1003 13:55:16.986173 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:55:16 crc kubenswrapper[4959]: I1003 13:55:16.988779 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.783218 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.785594 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.944258 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:55:17 crc kubenswrapper[4959]: E1003 13:55:17.944613 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="registry-server" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.944630 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="registry-server" Oct 03 13:55:17 crc kubenswrapper[4959]: E1003 13:55:17.944653 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="extract-utilities" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.944660 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="extract-utilities" Oct 03 13:55:17 crc kubenswrapper[4959]: E1003 13:55:17.944670 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="extract-content" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.944676 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="extract-content" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.944860 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7c8471-1c50-4862-83e6-eb27f7cdb788" containerName="registry-server" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.945805 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:17 crc kubenswrapper[4959]: I1003 13:55:17.956479 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.047904 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.047947 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.047977 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zbxp\" (UniqueName: \"kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.048167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.048294 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.048495 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150346 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150518 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zbxp\" (UniqueName: \"kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.150648 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.151507 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.151587 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.151659 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.151718 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.152118 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.174477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zbxp\" (UniqueName: \"kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp\") pod \"dnsmasq-dns-59cf4bdb65-fssm2\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.267845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:18 crc kubenswrapper[4959]: I1003 13:55:18.817735 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:55:18 crc kubenswrapper[4959]: W1003 13:55:18.851057 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb01d41b_854a_4072_8b05_807edc22812f.slice/crio-4e490e880c7b9fe405ba3fd81d5dcffec829e669e020897e0d7009a2412279aa WatchSource:0}: Error finding container 4e490e880c7b9fe405ba3fd81d5dcffec829e669e020897e0d7009a2412279aa: Status 404 returned error can't find the container with id 4e490e880c7b9fe405ba3fd81d5dcffec829e669e020897e0d7009a2412279aa Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.805171 4959 generic.go:334] "Generic (PLEG): container finished" podID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" containerID="286c3a4d76e172ce7cfa4c2cf6b0f1d71271dbefac5d465aa78b0eb9555090f9" exitCode=137 Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.805613 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b526ecc0-1b95-46d9-b02e-b1e688f5c969","Type":"ContainerDied","Data":"286c3a4d76e172ce7cfa4c2cf6b0f1d71271dbefac5d465aa78b0eb9555090f9"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.805637 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b526ecc0-1b95-46d9-b02e-b1e688f5c969","Type":"ContainerDied","Data":"b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.805647 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b87e93de9de42a5052c1d6fd4304a60065f2d987f6ce4805a235c42c4ebb5d8a" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.809047 4959 generic.go:334] "Generic (PLEG): container finished" podID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerID="f9a14d3f342cd387cfde7a61857031e7aba2a3acb904f84b2b45d694964778f3" exitCode=137 Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.809095 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerDied","Data":"f9a14d3f342cd387cfde7a61857031e7aba2a3acb904f84b2b45d694964778f3"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.809117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"21c3d548-ad08-44f0-a7b2-d8eeadd7d665","Type":"ContainerDied","Data":"dc4289968451daa1c0d66cb7db694f5a475d005820c4fe4157988248bc3cf3c4"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.809127 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc4289968451daa1c0d66cb7db694f5a475d005820c4fe4157988248bc3cf3c4" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.810259 4959 generic.go:334] "Generic (PLEG): container finished" podID="fb01d41b-854a-4072-8b05-807edc22812f" containerID="8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf" exitCode=0 Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.811570 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" event={"ID":"fb01d41b-854a-4072-8b05-807edc22812f","Type":"ContainerDied","Data":"8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.811597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" event={"ID":"fb01d41b-854a-4072-8b05-807edc22812f","Type":"ContainerStarted","Data":"4e490e880c7b9fe405ba3fd81d5dcffec829e669e020897e0d7009a2412279aa"} Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.882654 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.887225 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.982647 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx78n\" (UniqueName: \"kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n\") pod \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.982719 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle\") pod \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.982803 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle\") pod \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.982919 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs\") pod \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.982967 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k99c7\" (UniqueName: \"kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7\") pod \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.983023 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data\") pod \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\" (UID: \"b526ecc0-1b95-46d9-b02e-b1e688f5c969\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.983098 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data\") pod \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\" (UID: \"21c3d548-ad08-44f0-a7b2-d8eeadd7d665\") " Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.984471 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs" (OuterVolumeSpecName: "logs") pod "21c3d548-ad08-44f0-a7b2-d8eeadd7d665" (UID: "21c3d548-ad08-44f0-a7b2-d8eeadd7d665"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.985118 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.987427 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7" (OuterVolumeSpecName: "kube-api-access-k99c7") pod "21c3d548-ad08-44f0-a7b2-d8eeadd7d665" (UID: "21c3d548-ad08-44f0-a7b2-d8eeadd7d665"). InnerVolumeSpecName "kube-api-access-k99c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:19 crc kubenswrapper[4959]: I1003 13:55:19.988117 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n" (OuterVolumeSpecName: "kube-api-access-gx78n") pod "b526ecc0-1b95-46d9-b02e-b1e688f5c969" (UID: "b526ecc0-1b95-46d9-b02e-b1e688f5c969"). InnerVolumeSpecName "kube-api-access-gx78n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.018129 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21c3d548-ad08-44f0-a7b2-d8eeadd7d665" (UID: "21c3d548-ad08-44f0-a7b2-d8eeadd7d665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.019817 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data" (OuterVolumeSpecName: "config-data") pod "21c3d548-ad08-44f0-a7b2-d8eeadd7d665" (UID: "21c3d548-ad08-44f0-a7b2-d8eeadd7d665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.024338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data" (OuterVolumeSpecName: "config-data") pod "b526ecc0-1b95-46d9-b02e-b1e688f5c969" (UID: "b526ecc0-1b95-46d9-b02e-b1e688f5c969"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.028899 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b526ecc0-1b95-46d9-b02e-b1e688f5c969" (UID: "b526ecc0-1b95-46d9-b02e-b1e688f5c969"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090482 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090522 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx78n\" (UniqueName: \"kubernetes.io/projected/b526ecc0-1b95-46d9-b02e-b1e688f5c969-kube-api-access-gx78n\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090535 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090548 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090558 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k99c7\" (UniqueName: \"kubernetes.io/projected/21c3d548-ad08-44f0-a7b2-d8eeadd7d665-kube-api-access-k99c7\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.090570 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b526ecc0-1b95-46d9-b02e-b1e688f5c969-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.393479 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.393758 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-central-agent" containerID="cri-o://36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.393865 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="sg-core" containerID="cri-o://82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.393895 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-notification-agent" containerID="cri-o://18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.393881 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="proxy-httpd" containerID="cri-o://9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.494387 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.190:3000/\": read tcp 10.217.0.2:47174->10.217.0.190:3000: read: connection reset by peer" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834640 4959 generic.go:334] "Generic (PLEG): container finished" podID="11571651-7437-464a-8066-b4e1ae8c0774" containerID="9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af" exitCode=0 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834898 4959 generic.go:334] "Generic (PLEG): container finished" podID="11571651-7437-464a-8066-b4e1ae8c0774" containerID="82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956" exitCode=2 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834908 4959 generic.go:334] "Generic (PLEG): container finished" podID="11571651-7437-464a-8066-b4e1ae8c0774" containerID="36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c" exitCode=0 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834945 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerDied","Data":"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af"} Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerDied","Data":"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956"} Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.834998 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerDied","Data":"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c"} Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.840452 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" event={"ID":"fb01d41b-854a-4072-8b05-807edc22812f","Type":"ContainerStarted","Data":"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce"} Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.840482 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.840504 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.840829 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.842987 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.843178 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-log" containerID="cri-o://7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.843227 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-api" containerID="cri-o://0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6" gracePeriod=30 Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.895894 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" podStartSLOduration=3.895872248 podStartE2EDuration="3.895872248s" podCreationTimestamp="2025-10-03 13:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:20.872902759 +0000 UTC m=+1490.076246176" watchObservedRunningTime="2025-10-03 13:55:20.895872248 +0000 UTC m=+1490.099215665" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.908173 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.917987 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.931119 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: E1003 13:55:20.931650 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-log" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.931675 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-log" Oct 03 13:55:20 crc kubenswrapper[4959]: E1003 13:55:20.931694 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.931703 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:55:20 crc kubenswrapper[4959]: E1003 13:55:20.931715 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-metadata" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.931723 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-metadata" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.932143 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.932165 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-log" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.932184 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" containerName="nova-metadata-metadata" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.933056 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.937652 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.937670 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.937923 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.948687 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.961216 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.983246 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:55:20 crc kubenswrapper[4959]: I1003 13:55:20.998249 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:20.999987 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.001898 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.005150 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.029882 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108143 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108223 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108264 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84q6w\" (UniqueName: \"kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108287 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108390 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108461 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw9sb\" (UniqueName: \"kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.108512 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210004 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84q6w\" (UniqueName: \"kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210129 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210184 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210265 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw9sb\" (UniqueName: \"kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210332 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.210437 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.212043 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.216633 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.217856 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.217863 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.219500 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.220053 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.220645 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.225607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.231359 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw9sb\" (UniqueName: \"kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb\") pod \"nova-cell1-novncproxy-0\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.251095 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84q6w\" (UniqueName: \"kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w\") pod \"nova-metadata-0\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.317314 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.330345 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.701824 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c3d548-ad08-44f0-a7b2-d8eeadd7d665" path="/var/lib/kubelet/pods/21c3d548-ad08-44f0-a7b2-d8eeadd7d665/volumes" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.702691 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b526ecc0-1b95-46d9-b02e-b1e688f5c969" path="/var/lib/kubelet/pods/b526ecc0-1b95-46d9-b02e-b1e688f5c969/volumes" Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.825338 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.839864 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.850223 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerStarted","Data":"b83999b838fe333a68281fc60c8e645d6adb3230e44f3d32590e24545e2d7589"} Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.853182 4959 generic.go:334] "Generic (PLEG): container finished" podID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerID="7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8" exitCode=143 Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.853293 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerDied","Data":"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8"} Oct 03 13:55:21 crc kubenswrapper[4959]: I1003 13:55:21.855154 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"56e5d3a5-a733-402b-b393-d07706d30c48","Type":"ContainerStarted","Data":"cb13e6ee4cff2a3395b564828dea636a2d16c60883906eb1fff7b014b19f128d"} Oct 03 13:55:22 crc kubenswrapper[4959]: I1003 13:55:22.864341 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"56e5d3a5-a733-402b-b393-d07706d30c48","Type":"ContainerStarted","Data":"e9f6f6f4ded33babe3c6b4e9a7384c3513fe53da441505d10b5d13c79bd8152e"} Oct 03 13:55:22 crc kubenswrapper[4959]: I1003 13:55:22.866758 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerStarted","Data":"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1"} Oct 03 13:55:22 crc kubenswrapper[4959]: I1003 13:55:22.866830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerStarted","Data":"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55"} Oct 03 13:55:22 crc kubenswrapper[4959]: I1003 13:55:22.888718 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.88870144 podStartE2EDuration="2.88870144s" podCreationTimestamp="2025-10-03 13:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:22.882259933 +0000 UTC m=+1492.085603350" watchObservedRunningTime="2025-10-03 13:55:22.88870144 +0000 UTC m=+1492.092044857" Oct 03 13:55:22 crc kubenswrapper[4959]: I1003 13:55:22.907577 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.90755667 podStartE2EDuration="2.90755667s" podCreationTimestamp="2025-10-03 13:55:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:22.901126923 +0000 UTC m=+1492.104470340" watchObservedRunningTime="2025-10-03 13:55:22.90755667 +0000 UTC m=+1492.110900097" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.456513 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.568062 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data\") pod \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.568164 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9n9m\" (UniqueName: \"kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m\") pod \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.568374 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle\") pod \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.568442 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs\") pod \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\" (UID: \"5ec74ff6-4e1e-41c8-b347-2cf54add473d\") " Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.569606 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs" (OuterVolumeSpecName: "logs") pod "5ec74ff6-4e1e-41c8-b347-2cf54add473d" (UID: "5ec74ff6-4e1e-41c8-b347-2cf54add473d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.578299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m" (OuterVolumeSpecName: "kube-api-access-q9n9m") pod "5ec74ff6-4e1e-41c8-b347-2cf54add473d" (UID: "5ec74ff6-4e1e-41c8-b347-2cf54add473d"). InnerVolumeSpecName "kube-api-access-q9n9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.603158 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ec74ff6-4e1e-41c8-b347-2cf54add473d" (UID: "5ec74ff6-4e1e-41c8-b347-2cf54add473d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.610678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data" (OuterVolumeSpecName: "config-data") pod "5ec74ff6-4e1e-41c8-b347-2cf54add473d" (UID: "5ec74ff6-4e1e-41c8-b347-2cf54add473d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.671115 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.671165 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec74ff6-4e1e-41c8-b347-2cf54add473d-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.671178 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec74ff6-4e1e-41c8-b347-2cf54add473d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.671203 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9n9m\" (UniqueName: \"kubernetes.io/projected/5ec74ff6-4e1e-41c8-b347-2cf54add473d-kube-api-access-q9n9m\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.887737 4959 generic.go:334] "Generic (PLEG): container finished" podID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerID="0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6" exitCode=0 Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.887831 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.887827 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerDied","Data":"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6"} Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.888770 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5ec74ff6-4e1e-41c8-b347-2cf54add473d","Type":"ContainerDied","Data":"c636c885b3e171acb990705278895f26cd7bec04f0e2ab18c7d26c45808d3466"} Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.888803 4959 scope.go:117] "RemoveContainer" containerID="0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.918297 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.921738 4959 scope.go:117] "RemoveContainer" containerID="7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.930997 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.945016 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:24 crc kubenswrapper[4959]: E1003 13:55:24.945505 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-api" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.945523 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-api" Oct 03 13:55:24 crc kubenswrapper[4959]: E1003 13:55:24.945538 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-log" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.945544 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-log" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.945735 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-api" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.945753 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" containerName="nova-api-log" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.946742 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.950869 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.950904 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.950968 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.953156 4959 scope.go:117] "RemoveContainer" containerID="0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.957418 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:24 crc kubenswrapper[4959]: E1003 13:55:24.959895 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6\": container with ID starting with 0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6 not found: ID does not exist" containerID="0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.959938 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6"} err="failed to get container status \"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6\": rpc error: code = NotFound desc = could not find container \"0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6\": container with ID starting with 0baae9722ea9b85bf4db19f6e7ade92891689e908473c1e0e5d3178e6e5da5a6 not found: ID does not exist" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.959966 4959 scope.go:117] "RemoveContainer" containerID="7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8" Oct 03 13:55:24 crc kubenswrapper[4959]: E1003 13:55:24.964095 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8\": container with ID starting with 7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8 not found: ID does not exist" containerID="7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8" Oct 03 13:55:24 crc kubenswrapper[4959]: I1003 13:55:24.964132 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8"} err="failed to get container status \"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8\": rpc error: code = NotFound desc = could not find container \"7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8\": container with ID starting with 7eca6ed48d630a2eaa8a05db28cfe6c61e24ca62a3633e6da099a4be7b40b2e8 not found: ID does not exist" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.093345 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.093521 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.093748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.093795 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.093975 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgd9g\" (UniqueName: \"kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.094508 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.195834 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.195886 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.195939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgd9g\" (UniqueName: \"kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.196012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.196043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.196089 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.196731 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.199714 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.199897 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.200612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.235396 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.246825 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgd9g\" (UniqueName: \"kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g\") pod \"nova-api-0\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.274740 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.604126 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607393 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607430 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607453 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607471 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607519 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607555 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607675 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz5w8\" (UniqueName: \"kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8\") pod \"11571651-7437-464a-8066-b4e1ae8c0774\" (UID: \"11571651-7437-464a-8066-b4e1ae8c0774\") " Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.607952 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.608134 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.613179 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts" (OuterVolumeSpecName: "scripts") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.613824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8" (OuterVolumeSpecName: "kube-api-access-wz5w8") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "kube-api-access-wz5w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.651630 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.696837 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec74ff6-4e1e-41c8-b347-2cf54add473d" path="/var/lib/kubelet/pods/5ec74ff6-4e1e-41c8-b347-2cf54add473d/volumes" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.709825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710540 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz5w8\" (UniqueName: \"kubernetes.io/projected/11571651-7437-464a-8066-b4e1ae8c0774-kube-api-access-wz5w8\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710573 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710585 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710593 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11571651-7437-464a-8066-b4e1ae8c0774-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710602 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.710610 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.718483 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.730713 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data" (OuterVolumeSpecName: "config-data") pod "11571651-7437-464a-8066-b4e1ae8c0774" (UID: "11571651-7437-464a-8066-b4e1ae8c0774"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.754489 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:25 crc kubenswrapper[4959]: W1003 13:55:25.758094 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30c79fbc_e66b_420a_be52_c71720e24787.slice/crio-c11461d4613c12a951dadb5862306758321660bb8e9b5b550d2587edca6a75cb WatchSource:0}: Error finding container c11461d4613c12a951dadb5862306758321660bb8e9b5b550d2587edca6a75cb: Status 404 returned error can't find the container with id c11461d4613c12a951dadb5862306758321660bb8e9b5b550d2587edca6a75cb Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.812752 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.812826 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11571651-7437-464a-8066-b4e1ae8c0774-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.900684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerStarted","Data":"c11461d4613c12a951dadb5862306758321660bb8e9b5b550d2587edca6a75cb"} Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.904829 4959 generic.go:334] "Generic (PLEG): container finished" podID="11571651-7437-464a-8066-b4e1ae8c0774" containerID="18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666" exitCode=0 Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.904873 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerDied","Data":"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666"} Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.904901 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11571651-7437-464a-8066-b4e1ae8c0774","Type":"ContainerDied","Data":"85800a49b72e4df0d36e466e500bca16159ca593f2014e92de316822e8cd451e"} Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.904908 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.904918 4959 scope.go:117] "RemoveContainer" containerID="9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.926200 4959 scope.go:117] "RemoveContainer" containerID="82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.940142 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.949036 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.957921 4959 scope.go:117] "RemoveContainer" containerID="18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.963814 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:25 crc kubenswrapper[4959]: E1003 13:55:25.964663 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="sg-core" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.964686 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="sg-core" Oct 03 13:55:25 crc kubenswrapper[4959]: E1003 13:55:25.964705 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-notification-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.964713 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-notification-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: E1003 13:55:25.964728 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-central-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.964736 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-central-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: E1003 13:55:25.964748 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="proxy-httpd" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.964755 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="proxy-httpd" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.964981 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="sg-core" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.965014 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="proxy-httpd" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.965031 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-notification-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.965055 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="11571651-7437-464a-8066-b4e1ae8c0774" containerName="ceilometer-central-agent" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.967149 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.970448 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.970532 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.970594 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:55:25 crc kubenswrapper[4959]: I1003 13:55:25.987245 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.010622 4959 scope.go:117] "RemoveContainer" containerID="36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.036636 4959 scope.go:117] "RemoveContainer" containerID="9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af" Oct 03 13:55:26 crc kubenswrapper[4959]: E1003 13:55:26.037053 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af\": container with ID starting with 9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af not found: ID does not exist" containerID="9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037087 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af"} err="failed to get container status \"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af\": rpc error: code = NotFound desc = could not find container \"9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af\": container with ID starting with 9e61fe3c121ed5d0799a1453b53350e1aa71a8aead75dfdf03d472f58ebfd1af not found: ID does not exist" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037107 4959 scope.go:117] "RemoveContainer" containerID="82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956" Oct 03 13:55:26 crc kubenswrapper[4959]: E1003 13:55:26.037497 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956\": container with ID starting with 82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956 not found: ID does not exist" containerID="82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037546 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956"} err="failed to get container status \"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956\": rpc error: code = NotFound desc = could not find container \"82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956\": container with ID starting with 82e303d5205e2cf3b4a63aa97a4bf26d6154a0aa7e0a0dd0a74247417fa73956 not found: ID does not exist" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037589 4959 scope.go:117] "RemoveContainer" containerID="18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666" Oct 03 13:55:26 crc kubenswrapper[4959]: E1003 13:55:26.037917 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666\": container with ID starting with 18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666 not found: ID does not exist" containerID="18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037940 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666"} err="failed to get container status \"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666\": rpc error: code = NotFound desc = could not find container \"18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666\": container with ID starting with 18823526cd359579002d171b3615f1ed1aca9882b4c7e7d7349325f0a6ca2666 not found: ID does not exist" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.037954 4959 scope.go:117] "RemoveContainer" containerID="36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c" Oct 03 13:55:26 crc kubenswrapper[4959]: E1003 13:55:26.038135 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c\": container with ID starting with 36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c not found: ID does not exist" containerID="36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.038336 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c"} err="failed to get container status \"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c\": rpc error: code = NotFound desc = could not find container \"36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c\": container with ID starting with 36664735d000becf9ad8fc9f48298990dcebb7cea6f78a581bb9ce005e9cc54c not found: ID does not exist" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117586 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117671 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117741 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117922 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117943 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.117970 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmdvz\" (UniqueName: \"kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220429 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220583 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmdvz\" (UniqueName: \"kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220675 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220773 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220876 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.220977 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.221159 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.221324 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.223996 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.224929 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.225685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.226047 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.227227 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.236918 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmdvz\" (UniqueName: \"kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz\") pod \"ceilometer-0\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.296576 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.317894 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.330811 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.330860 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.736658 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.914437 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerStarted","Data":"008fd636ed07ff2229a285a6c6a34162f656b78166e80c2b35d0f103b465bcd6"} Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.916515 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerStarted","Data":"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93"} Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.916564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerStarted","Data":"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486"} Oct 03 13:55:26 crc kubenswrapper[4959]: I1003 13:55:26.939650 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.939627055 podStartE2EDuration="2.939627055s" podCreationTimestamp="2025-10-03 13:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:26.934667313 +0000 UTC m=+1496.138010750" watchObservedRunningTime="2025-10-03 13:55:26.939627055 +0000 UTC m=+1496.142970472" Oct 03 13:55:27 crc kubenswrapper[4959]: I1003 13:55:27.698216 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11571651-7437-464a-8066-b4e1ae8c0774" path="/var/lib/kubelet/pods/11571651-7437-464a-8066-b4e1ae8c0774/volumes" Oct 03 13:55:27 crc kubenswrapper[4959]: I1003 13:55:27.929769 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerStarted","Data":"4ddb932943c117d08c1006b64f264db30d2dae0973fb4d5017d89ad84b3ec3a8"} Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.269253 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.331219 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.331511 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="dnsmasq-dns" containerID="cri-o://80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f" gracePeriod=10 Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.913730 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.948342 4959 generic.go:334] "Generic (PLEG): container finished" podID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerID="80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f" exitCode=0 Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.948404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" event={"ID":"5f6e101d-a3cf-4573-8c49-a8b81b9f0766","Type":"ContainerDied","Data":"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f"} Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.948428 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" event={"ID":"5f6e101d-a3cf-4573-8c49-a8b81b9f0766","Type":"ContainerDied","Data":"f5b239987c348b417302d77a22d43946766f904e48ba70d16d9cb0d5fe4156fb"} Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.948444 4959 scope.go:117] "RemoveContainer" containerID="80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f" Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.948550 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hstfl" Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.961108 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerStarted","Data":"47160c83891466f0f38183ffc99e4cb435c5e8b448cf8d1325e15334ebeb44d0"} Oct 03 13:55:28 crc kubenswrapper[4959]: I1003 13:55:28.983802 4959 scope.go:117] "RemoveContainer" containerID="6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.010752 4959 scope.go:117] "RemoveContainer" containerID="80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f" Oct 03 13:55:29 crc kubenswrapper[4959]: E1003 13:55:29.011620 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f\": container with ID starting with 80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f not found: ID does not exist" containerID="80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.011646 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f"} err="failed to get container status \"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f\": rpc error: code = NotFound desc = could not find container \"80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f\": container with ID starting with 80fd4cba54843a6fac0afb3e85b0f6043457b8010643c9f3da38b8863e0cff9f not found: ID does not exist" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.011667 4959 scope.go:117] "RemoveContainer" containerID="6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4" Oct 03 13:55:29 crc kubenswrapper[4959]: E1003 13:55:29.015934 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4\": container with ID starting with 6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4 not found: ID does not exist" containerID="6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.015990 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4"} err="failed to get container status \"6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4\": rpc error: code = NotFound desc = could not find container \"6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4\": container with ID starting with 6281d6c9ffa3295ca87afe1b281c5b24365ce6ff0f33cace7d8780886d8284f4 not found: ID does not exist" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079338 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079362 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079383 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079416 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zd2p\" (UniqueName: \"kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.079534 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb\") pod \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\" (UID: \"5f6e101d-a3cf-4573-8c49-a8b81b9f0766\") " Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.084893 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p" (OuterVolumeSpecName: "kube-api-access-5zd2p") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "kube-api-access-5zd2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.129618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.133675 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.133990 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.139523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.140387 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config" (OuterVolumeSpecName: "config") pod "5f6e101d-a3cf-4573-8c49-a8b81b9f0766" (UID: "5f6e101d-a3cf-4573-8c49-a8b81b9f0766"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181263 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181299 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181307 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181318 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181327 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.181336 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zd2p\" (UniqueName: \"kubernetes.io/projected/5f6e101d-a3cf-4573-8c49-a8b81b9f0766-kube-api-access-5zd2p\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.290788 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.299689 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hstfl"] Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.698614 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" path="/var/lib/kubelet/pods/5f6e101d-a3cf-4573-8c49-a8b81b9f0766/volumes" Oct 03 13:55:29 crc kubenswrapper[4959]: I1003 13:55:29.973202 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerStarted","Data":"13ffbebcbaac00743c439cf6bbe0144cf9261560d362f82d27b626d91213b204"} Oct 03 13:55:30 crc kubenswrapper[4959]: I1003 13:55:30.983769 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerStarted","Data":"ada15b2c15fe2f2dacc57422fd81cb3d234b56614971b21c30ab71da932c7b32"} Oct 03 13:55:30 crc kubenswrapper[4959]: I1003 13:55:30.984271 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:55:31 crc kubenswrapper[4959]: I1003 13:55:31.011570 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.43142252 podStartE2EDuration="6.011543881s" podCreationTimestamp="2025-10-03 13:55:25 +0000 UTC" firstStartedPulling="2025-10-03 13:55:26.74268282 +0000 UTC m=+1495.946026237" lastFinishedPulling="2025-10-03 13:55:30.322804181 +0000 UTC m=+1499.526147598" observedRunningTime="2025-10-03 13:55:30.999745573 +0000 UTC m=+1500.203089010" watchObservedRunningTime="2025-10-03 13:55:31.011543881 +0000 UTC m=+1500.214887318" Oct 03 13:55:31 crc kubenswrapper[4959]: I1003 13:55:31.318882 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:31 crc kubenswrapper[4959]: I1003 13:55:31.330797 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:55:31 crc kubenswrapper[4959]: I1003 13:55:31.330892 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:55:31 crc kubenswrapper[4959]: I1003 13:55:31.337141 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.024614 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.200270 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-55hsl"] Oct 03 13:55:32 crc kubenswrapper[4959]: E1003 13:55:32.200661 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="dnsmasq-dns" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.200684 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="dnsmasq-dns" Oct 03 13:55:32 crc kubenswrapper[4959]: E1003 13:55:32.200727 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="init" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.200735 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="init" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.200957 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6e101d-a3cf-4573-8c49-a8b81b9f0766" containerName="dnsmasq-dns" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.202071 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.204070 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.207690 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.220688 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-55hsl"] Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.244990 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.245101 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.245224 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9pvt\" (UniqueName: \"kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.245326 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.346875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.347336 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.347430 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.347459 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9pvt\" (UniqueName: \"kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.347533 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.347742 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.354313 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.355843 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.356935 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.366649 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9pvt\" (UniqueName: \"kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt\") pod \"nova-cell1-cell-mapping-55hsl\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.519706 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:32 crc kubenswrapper[4959]: I1003 13:55:32.992048 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-55hsl"] Oct 03 13:55:33 crc kubenswrapper[4959]: W1003 13:55:33.001405 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2b4233a_0bb4_41bc_a81f_1a299aa9308d.slice/crio-9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7 WatchSource:0}: Error finding container 9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7: Status 404 returned error can't find the container with id 9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7 Oct 03 13:55:33 crc kubenswrapper[4959]: I1003 13:55:33.021439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-55hsl" event={"ID":"e2b4233a-0bb4-41bc-a81f-1a299aa9308d","Type":"ContainerStarted","Data":"9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7"} Oct 03 13:55:34 crc kubenswrapper[4959]: I1003 13:55:34.034244 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-55hsl" event={"ID":"e2b4233a-0bb4-41bc-a81f-1a299aa9308d","Type":"ContainerStarted","Data":"8c342ba7a12a3d1b866f17f84e61cfc1269c690927daba2bc2a8e0c10ed68b3c"} Oct 03 13:55:34 crc kubenswrapper[4959]: I1003 13:55:34.073833 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-55hsl" podStartSLOduration=2.073804479 podStartE2EDuration="2.073804479s" podCreationTimestamp="2025-10-03 13:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:34.069463513 +0000 UTC m=+1503.272806950" watchObservedRunningTime="2025-10-03 13:55:34.073804479 +0000 UTC m=+1503.277147936" Oct 03 13:55:35 crc kubenswrapper[4959]: I1003 13:55:35.276005 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:35 crc kubenswrapper[4959]: I1003 13:55:35.276519 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:36 crc kubenswrapper[4959]: I1003 13:55:36.044114 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:55:36 crc kubenswrapper[4959]: I1003 13:55:36.044633 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:55:36 crc kubenswrapper[4959]: I1003 13:55:36.288448 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:36 crc kubenswrapper[4959]: I1003 13:55:36.289550 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:38 crc kubenswrapper[4959]: I1003 13:55:38.072880 4959 generic.go:334] "Generic (PLEG): container finished" podID="e2b4233a-0bb4-41bc-a81f-1a299aa9308d" containerID="8c342ba7a12a3d1b866f17f84e61cfc1269c690927daba2bc2a8e0c10ed68b3c" exitCode=0 Oct 03 13:55:38 crc kubenswrapper[4959]: I1003 13:55:38.073054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-55hsl" event={"ID":"e2b4233a-0bb4-41bc-a81f-1a299aa9308d","Type":"ContainerDied","Data":"8c342ba7a12a3d1b866f17f84e61cfc1269c690927daba2bc2a8e0c10ed68b3c"} Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.458300 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.516779 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle\") pod \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.516868 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data\") pod \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.516894 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9pvt\" (UniqueName: \"kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt\") pod \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.517015 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts\") pod \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\" (UID: \"e2b4233a-0bb4-41bc-a81f-1a299aa9308d\") " Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.523095 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt" (OuterVolumeSpecName: "kube-api-access-t9pvt") pod "e2b4233a-0bb4-41bc-a81f-1a299aa9308d" (UID: "e2b4233a-0bb4-41bc-a81f-1a299aa9308d"). InnerVolumeSpecName "kube-api-access-t9pvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.541441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts" (OuterVolumeSpecName: "scripts") pod "e2b4233a-0bb4-41bc-a81f-1a299aa9308d" (UID: "e2b4233a-0bb4-41bc-a81f-1a299aa9308d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.553821 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data" (OuterVolumeSpecName: "config-data") pod "e2b4233a-0bb4-41bc-a81f-1a299aa9308d" (UID: "e2b4233a-0bb4-41bc-a81f-1a299aa9308d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.562116 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2b4233a-0bb4-41bc-a81f-1a299aa9308d" (UID: "e2b4233a-0bb4-41bc-a81f-1a299aa9308d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.619770 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.619803 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.619813 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9pvt\" (UniqueName: \"kubernetes.io/projected/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-kube-api-access-t9pvt\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:39 crc kubenswrapper[4959]: I1003 13:55:39.619823 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2b4233a-0bb4-41bc-a81f-1a299aa9308d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.101796 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-55hsl" event={"ID":"e2b4233a-0bb4-41bc-a81f-1a299aa9308d","Type":"ContainerDied","Data":"9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7"} Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.102046 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a230429fd10a89e9a563962273cb73a254457fe10892b2af5d20aeb28478fc7" Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.101896 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-55hsl" Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.312160 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.312517 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerName="nova-scheduler-scheduler" containerID="cri-o://ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" gracePeriod=30 Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.331038 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.331355 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-log" containerID="cri-o://4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486" gracePeriod=30 Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.331411 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-api" containerID="cri-o://596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93" gracePeriod=30 Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.349899 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.350132 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-log" containerID="cri-o://3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55" gracePeriod=30 Oct 03 13:55:40 crc kubenswrapper[4959]: I1003 13:55:40.350240 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-metadata" containerID="cri-o://b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1" gracePeriod=30 Oct 03 13:55:41 crc kubenswrapper[4959]: I1003 13:55:41.110608 4959 generic.go:334] "Generic (PLEG): container finished" podID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerID="3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55" exitCode=143 Oct 03 13:55:41 crc kubenswrapper[4959]: I1003 13:55:41.110667 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerDied","Data":"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55"} Oct 03 13:55:41 crc kubenswrapper[4959]: I1003 13:55:41.112329 4959 generic.go:334] "Generic (PLEG): container finished" podID="30c79fbc-e66b-420a-be52-c71720e24787" containerID="4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486" exitCode=143 Oct 03 13:55:41 crc kubenswrapper[4959]: I1003 13:55:41.112352 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerDied","Data":"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486"} Oct 03 13:55:41 crc kubenswrapper[4959]: E1003 13:55:41.984721 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:55:41 crc kubenswrapper[4959]: E1003 13:55:41.986797 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:55:41 crc kubenswrapper[4959]: E1003 13:55:41.988382 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:55:41 crc kubenswrapper[4959]: E1003 13:55:41.988445 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerName="nova-scheduler-scheduler" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.056736 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.068542 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.102916 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103031 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84q6w\" (UniqueName: \"kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w\") pod \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103111 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103171 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs\") pod \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103221 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle\") pod \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103255 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs\") pod \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103343 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgd9g\" (UniqueName: \"kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103368 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103442 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle\") pod \"30c79fbc-e66b-420a-be52-c71720e24787\" (UID: \"30c79fbc-e66b-420a-be52-c71720e24787\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.103620 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data\") pod \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\" (UID: \"b49b3b7b-5c67-43dd-a37c-fe30d617421c\") " Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.113543 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs" (OuterVolumeSpecName: "logs") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.113566 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs" (OuterVolumeSpecName: "logs") pod "b49b3b7b-5c67-43dd-a37c-fe30d617421c" (UID: "b49b3b7b-5c67-43dd-a37c-fe30d617421c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.127520 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w" (OuterVolumeSpecName: "kube-api-access-84q6w") pod "b49b3b7b-5c67-43dd-a37c-fe30d617421c" (UID: "b49b3b7b-5c67-43dd-a37c-fe30d617421c"). InnerVolumeSpecName "kube-api-access-84q6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.127966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g" (OuterVolumeSpecName: "kube-api-access-vgd9g") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "kube-api-access-vgd9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.143871 4959 generic.go:334] "Generic (PLEG): container finished" podID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerID="b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1" exitCode=0 Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.143951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerDied","Data":"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1"} Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.143962 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.143984 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b49b3b7b-5c67-43dd-a37c-fe30d617421c","Type":"ContainerDied","Data":"b83999b838fe333a68281fc60c8e645d6adb3230e44f3d32590e24545e2d7589"} Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.144006 4959 scope.go:117] "RemoveContainer" containerID="b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.147376 4959 generic.go:334] "Generic (PLEG): container finished" podID="30c79fbc-e66b-420a-be52-c71720e24787" containerID="596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93" exitCode=0 Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.147413 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerDied","Data":"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93"} Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.147438 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30c79fbc-e66b-420a-be52-c71720e24787","Type":"ContainerDied","Data":"c11461d4613c12a951dadb5862306758321660bb8e9b5b550d2587edca6a75cb"} Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.147510 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.151054 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data" (OuterVolumeSpecName: "config-data") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.153555 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b49b3b7b-5c67-43dd-a37c-fe30d617421c" (UID: "b49b3b7b-5c67-43dd-a37c-fe30d617421c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.159597 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data" (OuterVolumeSpecName: "config-data") pod "b49b3b7b-5c67-43dd-a37c-fe30d617421c" (UID: "b49b3b7b-5c67-43dd-a37c-fe30d617421c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.177634 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.182673 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.184677 4959 scope.go:117] "RemoveContainer" containerID="3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.195301 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b49b3b7b-5c67-43dd-a37c-fe30d617421c" (UID: "b49b3b7b-5c67-43dd-a37c-fe30d617421c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.198357 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "30c79fbc-e66b-420a-be52-c71720e24787" (UID: "30c79fbc-e66b-420a-be52-c71720e24787"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.205774 4959 scope.go:117] "RemoveContainer" containerID="b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.206241 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1\": container with ID starting with b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1 not found: ID does not exist" containerID="b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.206300 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1"} err="failed to get container status \"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1\": rpc error: code = NotFound desc = could not find container \"b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1\": container with ID starting with b2e11dda65c0edd83a31f9350e02c17c6685e1aed6d78bcc890c006d9a41ece1 not found: ID does not exist" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.206334 4959 scope.go:117] "RemoveContainer" containerID="3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.206628 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55\": container with ID starting with 3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55 not found: ID does not exist" containerID="3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.206661 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55"} err="failed to get container status \"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55\": rpc error: code = NotFound desc = could not find container \"3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55\": container with ID starting with 3d16d7a22cc042d30c73a1344261896aea88c433a4a7165524b13b15bdd18d55 not found: ID does not exist" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.206683 4959 scope.go:117] "RemoveContainer" containerID="596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207341 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207384 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207398 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207410 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30c79fbc-e66b-420a-be52-c71720e24787-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207424 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84q6w\" (UniqueName: \"kubernetes.io/projected/b49b3b7b-5c67-43dd-a37c-fe30d617421c-kube-api-access-84q6w\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207438 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207449 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207461 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b49b3b7b-5c67-43dd-a37c-fe30d617421c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207471 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b49b3b7b-5c67-43dd-a37c-fe30d617421c-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207482 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgd9g\" (UniqueName: \"kubernetes.io/projected/30c79fbc-e66b-420a-be52-c71720e24787-kube-api-access-vgd9g\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.207493 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30c79fbc-e66b-420a-be52-c71720e24787-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.227431 4959 scope.go:117] "RemoveContainer" containerID="4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.247804 4959 scope.go:117] "RemoveContainer" containerID="596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.248279 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93\": container with ID starting with 596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93 not found: ID does not exist" containerID="596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.248321 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93"} err="failed to get container status \"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93\": rpc error: code = NotFound desc = could not find container \"596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93\": container with ID starting with 596ecdf0a4c2b6bc214fb7bdcf7ac3274c13473f6957a3d6b90e0fb659da5a93 not found: ID does not exist" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.248354 4959 scope.go:117] "RemoveContainer" containerID="4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.248859 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486\": container with ID starting with 4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486 not found: ID does not exist" containerID="4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.249258 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486"} err="failed to get container status \"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486\": rpc error: code = NotFound desc = could not find container \"4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486\": container with ID starting with 4b4782b16d796112a91d84ed24b2b8e60ee2fdfdc148db9cd6349004edad9486 not found: ID does not exist" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.499949 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.512041 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.531444 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.543486 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573242 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.573767 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-log" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573791 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-log" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.573814 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-log" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573825 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-log" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.573849 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b4233a-0bb4-41bc-a81f-1a299aa9308d" containerName="nova-manage" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573861 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b4233a-0bb4-41bc-a81f-1a299aa9308d" containerName="nova-manage" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.573888 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-metadata" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573898 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-metadata" Oct 03 13:55:44 crc kubenswrapper[4959]: E1003 13:55:44.573921 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-api" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.573929 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-api" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.574150 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-metadata" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.574177 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-log" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.574213 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b4233a-0bb4-41bc-a81f-1a299aa9308d" containerName="nova-manage" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.574229 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c79fbc-e66b-420a-be52-c71720e24787" containerName="nova-api-api" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.574249 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" containerName="nova-metadata-log" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.575522 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.579439 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.579690 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.579969 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.591296 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.593639 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.598478 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.599343 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.607575 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.617438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.617515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.618117 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.618202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.618263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.618386 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d5wz\" (UniqueName: \"kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.624985 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh78d\" (UniqueName: \"kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719670 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719785 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719810 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719842 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d5wz\" (UniqueName: \"kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.719969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.720874 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.725377 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.725409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.739029 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.739563 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.751259 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d5wz\" (UniqueName: \"kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz\") pod \"nova-api-0\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.821976 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.822262 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.822314 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.822337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.822501 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh78d\" (UniqueName: \"kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.823322 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.826510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.826636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.827291 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.839609 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh78d\" (UniqueName: \"kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d\") pod \"nova-metadata-0\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " pod="openstack/nova-metadata-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.941988 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:55:44 crc kubenswrapper[4959]: I1003 13:55:44.948047 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:55:45 crc kubenswrapper[4959]: I1003 13:55:45.424652 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:55:45 crc kubenswrapper[4959]: W1003 13:55:45.430054 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd375128e_6fb8_4c35_9def_4bd67c206d39.slice/crio-e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28 WatchSource:0}: Error finding container e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28: Status 404 returned error can't find the container with id e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28 Oct 03 13:55:45 crc kubenswrapper[4959]: I1003 13:55:45.494324 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:55:45 crc kubenswrapper[4959]: I1003 13:55:45.695124 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c79fbc-e66b-420a-be52-c71720e24787" path="/var/lib/kubelet/pods/30c79fbc-e66b-420a-be52-c71720e24787/volumes" Oct 03 13:55:45 crc kubenswrapper[4959]: I1003 13:55:45.696360 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b49b3b7b-5c67-43dd-a37c-fe30d617421c" path="/var/lib/kubelet/pods/b49b3b7b-5c67-43dd-a37c-fe30d617421c/volumes" Oct 03 13:55:45 crc kubenswrapper[4959]: I1003 13:55:45.968854 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.045412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data\") pod \"c5659216-fc65-4dbd-a13b-addef38f4ebd\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.045834 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khf66\" (UniqueName: \"kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66\") pod \"c5659216-fc65-4dbd-a13b-addef38f4ebd\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.045878 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle\") pod \"c5659216-fc65-4dbd-a13b-addef38f4ebd\" (UID: \"c5659216-fc65-4dbd-a13b-addef38f4ebd\") " Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.051440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66" (OuterVolumeSpecName: "kube-api-access-khf66") pod "c5659216-fc65-4dbd-a13b-addef38f4ebd" (UID: "c5659216-fc65-4dbd-a13b-addef38f4ebd"). InnerVolumeSpecName "kube-api-access-khf66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.072792 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5659216-fc65-4dbd-a13b-addef38f4ebd" (UID: "c5659216-fc65-4dbd-a13b-addef38f4ebd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.078700 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data" (OuterVolumeSpecName: "config-data") pod "c5659216-fc65-4dbd-a13b-addef38f4ebd" (UID: "c5659216-fc65-4dbd-a13b-addef38f4ebd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.149016 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.149090 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khf66\" (UniqueName: \"kubernetes.io/projected/c5659216-fc65-4dbd-a13b-addef38f4ebd-kube-api-access-khf66\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.149119 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5659216-fc65-4dbd-a13b-addef38f4ebd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.175620 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerStarted","Data":"72a3e7929b0adc554a8bfd78df9f7708f0538f33a7bbf4518d4a306cb7e072a8"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.175692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerStarted","Data":"9f1517df64486c3bcc421232da8688f5478dc78c9cfeec76b3735eadbe8fbb25"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.175714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerStarted","Data":"e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.177717 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerStarted","Data":"cfb72705ebccca6e3af246841cf201b6036478e1d451c879eb40ab8cf65d5799"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.177751 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerStarted","Data":"bd8aaedeb008acc8fdb85e7ff1a53c8c9fefc46446b823e52046fd7ccab71695"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.177761 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerStarted","Data":"caae51897f61d45af2f7174fb8c8b4b34f4a757768c222ebc005f304dfb827c6"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.205742 4959 generic.go:334] "Generic (PLEG): container finished" podID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" exitCode=0 Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.206927 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.206911044 podStartE2EDuration="2.206911044s" podCreationTimestamp="2025-10-03 13:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:46.20594031 +0000 UTC m=+1515.409283747" watchObservedRunningTime="2025-10-03 13:55:46.206911044 +0000 UTC m=+1515.410254461" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.207772 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.205818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5659216-fc65-4dbd-a13b-addef38f4ebd","Type":"ContainerDied","Data":"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.208182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c5659216-fc65-4dbd-a13b-addef38f4ebd","Type":"ContainerDied","Data":"b4aaa741d1e8ac948c3df7d7539cd5c5917034c6f3244fa3d8d1ec7665a9b5ff"} Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.208345 4959 scope.go:117] "RemoveContainer" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.234631 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.234613209 podStartE2EDuration="2.234613209s" podCreationTimestamp="2025-10-03 13:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:46.232718013 +0000 UTC m=+1515.436061430" watchObservedRunningTime="2025-10-03 13:55:46.234613209 +0000 UTC m=+1515.437956626" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.244252 4959 scope.go:117] "RemoveContainer" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" Oct 03 13:55:46 crc kubenswrapper[4959]: E1003 13:55:46.244909 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7\": container with ID starting with ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7 not found: ID does not exist" containerID="ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.244951 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7"} err="failed to get container status \"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7\": rpc error: code = NotFound desc = could not find container \"ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7\": container with ID starting with ba419b74328de4426da48764f562b0a45186c111e30bbe3de82746a4aa8b9eb7 not found: ID does not exist" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.259554 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.271162 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.285628 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:46 crc kubenswrapper[4959]: E1003 13:55:46.286063 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerName="nova-scheduler-scheduler" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.286081 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerName="nova-scheduler-scheduler" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.286259 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" containerName="nova-scheduler-scheduler" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.286920 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.289133 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.293231 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.353122 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.353223 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.353253 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfwd7\" (UniqueName: \"kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.454803 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.454865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfwd7\" (UniqueName: \"kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.454969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.458533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.458962 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.472018 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfwd7\" (UniqueName: \"kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7\") pod \"nova-scheduler-0\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " pod="openstack/nova-scheduler-0" Oct 03 13:55:46 crc kubenswrapper[4959]: I1003 13:55:46.634804 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:55:47 crc kubenswrapper[4959]: I1003 13:55:47.190619 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:55:47 crc kubenswrapper[4959]: W1003 13:55:47.196024 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632a7ace_4d5d_4588_af93_e0ab25125a14.slice/crio-e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b WatchSource:0}: Error finding container e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b: Status 404 returned error can't find the container with id e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b Oct 03 13:55:47 crc kubenswrapper[4959]: I1003 13:55:47.217900 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"632a7ace-4d5d-4588-af93-e0ab25125a14","Type":"ContainerStarted","Data":"e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b"} Oct 03 13:55:47 crc kubenswrapper[4959]: I1003 13:55:47.695971 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5659216-fc65-4dbd-a13b-addef38f4ebd" path="/var/lib/kubelet/pods/c5659216-fc65-4dbd-a13b-addef38f4ebd/volumes" Oct 03 13:55:48 crc kubenswrapper[4959]: I1003 13:55:48.231328 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"632a7ace-4d5d-4588-af93-e0ab25125a14","Type":"ContainerStarted","Data":"1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a"} Oct 03 13:55:48 crc kubenswrapper[4959]: I1003 13:55:48.271796 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.2717770010000002 podStartE2EDuration="2.271777001s" podCreationTimestamp="2025-10-03 13:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:55:48.269315652 +0000 UTC m=+1517.472659079" watchObservedRunningTime="2025-10-03 13:55:48.271777001 +0000 UTC m=+1517.475120418" Oct 03 13:55:49 crc kubenswrapper[4959]: I1003 13:55:49.949066 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:55:49 crc kubenswrapper[4959]: I1003 13:55:49.949504 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:55:51 crc kubenswrapper[4959]: I1003 13:55:51.636003 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:55:54 crc kubenswrapper[4959]: I1003 13:55:54.942527 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:54 crc kubenswrapper[4959]: I1003 13:55:54.943137 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:55:54 crc kubenswrapper[4959]: I1003 13:55:54.948711 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:55:54 crc kubenswrapper[4959]: I1003 13:55:54.949082 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:55:55 crc kubenswrapper[4959]: I1003 13:55:55.958437 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:55 crc kubenswrapper[4959]: I1003 13:55:55.958855 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:55 crc kubenswrapper[4959]: I1003 13:55:55.976417 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:55 crc kubenswrapper[4959]: I1003 13:55:55.976422 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:55:56 crc kubenswrapper[4959]: I1003 13:55:56.317089 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 13:55:56 crc kubenswrapper[4959]: I1003 13:55:56.635386 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:55:56 crc kubenswrapper[4959]: I1003 13:55:56.670453 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:55:57 crc kubenswrapper[4959]: I1003 13:55:57.357346 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.956822 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.957716 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.958918 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.961478 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.964977 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.972291 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:56:04 crc kubenswrapper[4959]: I1003 13:56:04.973433 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:56:05 crc kubenswrapper[4959]: I1003 13:56:05.412509 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:56:05 crc kubenswrapper[4959]: I1003 13:56:05.416593 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:56:05 crc kubenswrapper[4959]: I1003 13:56:05.418536 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.044929 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.045001 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.045052 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.045901 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.045969 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" gracePeriod=600 Oct 03 13:56:06 crc kubenswrapper[4959]: E1003 13:56:06.197861 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.423945 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" exitCode=0 Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.423988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c"} Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.424042 4959 scope.go:117] "RemoveContainer" containerID="0fe62be3d52b17e7ad94c03b4cf7bdf6aa842e3ad290798a9796729c3fc84b77" Oct 03 13:56:06 crc kubenswrapper[4959]: I1003 13:56:06.425229 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:56:06 crc kubenswrapper[4959]: E1003 13:56:06.425528 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:56:20 crc kubenswrapper[4959]: I1003 13:56:20.686273 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:56:20 crc kubenswrapper[4959]: E1003 13:56:20.687529 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.172275 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.173072 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="a7088287-7362-42f5-987c-be9f700c50cb" containerName="openstackclient" containerID="cri-o://bae04dd3c4a3a80e408d2aa1da0eb1a0f82c686e02752a0f01f6986ff0f0542b" gracePeriod=2 Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.195290 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.225891 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:56:25 crc kubenswrapper[4959]: E1003 13:56:25.226321 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7088287-7362-42f5-987c-be9f700c50cb" containerName="openstackclient" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.226339 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7088287-7362-42f5-987c-be9f700c50cb" containerName="openstackclient" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.226545 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7088287-7362-42f5-987c-be9f700c50cb" containerName="openstackclient" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.232406 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.245708 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.247761 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.290038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.314271 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.329315 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.329375 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.329424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhcsd\" (UniqueName: \"kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.329500 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.329548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.410911 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.430695 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.431808 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.431862 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.431926 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.431961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.431994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432049 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432077 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhcsd\" (UniqueName: \"kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432135 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432173 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6c7g\" (UniqueName: \"kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432236 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432358 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.432407 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.451040 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.453660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.481821 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.486086 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhcsd\" (UniqueName: \"kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd\") pod \"barbican-keystone-listener-68677546bb-dgqv8\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.515963 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6c7g\" (UniqueName: \"kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534565 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534598 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534613 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckzln\" (UniqueName: \"kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534725 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534755 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534804 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534834 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.534851 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.536265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: E1003 13:56:25.538843 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 03 13:56:25 crc kubenswrapper[4959]: E1003 13:56:25.538904 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data podName:d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2 nodeName:}" failed. No retries permitted until 2025-10-03 13:56:26.038881788 +0000 UTC m=+1555.242225285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data") pod "rabbitmq-server-0" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2") : configmap "rabbitmq-config-data" not found Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.549770 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.561584 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.562201 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.565429 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.584823 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.587858 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.613950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6c7g\" (UniqueName: \"kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g\") pod \"barbican-worker-b7cd665dc-9szsx\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641160 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641197 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641236 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641280 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641328 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86czq\" (UniqueName: \"kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq\") pod \"placementaa19-account-delete-twnzm\" (UID: \"943c9966-87bc-44fb-a19f-4ae15db1bdc4\") " pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641369 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.641391 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckzln\" (UniqueName: \"kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.651859 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.660523 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.660870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.663936 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.665748 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.668413 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.675348 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.698643 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckzln\" (UniqueName: \"kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln\") pod \"barbican-api-544d968d58-6cpgj\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.733122 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.752690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86czq\" (UniqueName: \"kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq\") pod \"placementaa19-account-delete-twnzm\" (UID: \"943c9966-87bc-44fb-a19f-4ae15db1bdc4\") " pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.774210 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.775336 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.775443 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.803539 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.803954 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="openstack-network-exporter" containerID="cri-o://eba109515bfe8301850c53df177bc5231193d427a4c857c848ed3c8d12e4e21b" gracePeriod=300 Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.853954 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jnfg\" (UniqueName: \"kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg\") pod \"glance6565-account-delete-q7tmg\" (UID: \"1da07093-5ea1-464a-a806-d30494ae9929\") " pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:25 crc kubenswrapper[4959]: E1003 13:56:25.855055 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:25 crc kubenswrapper[4959]: E1003 13:56:25.855100 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:26.355086792 +0000 UTC m=+1555.558430209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.866913 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.867501 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="openstack-network-exporter" containerID="cri-o://18ea0e8b0cc153504a9c772e75f38038e8891d7353b017f050ebb59b0ecbf766" gracePeriod=300 Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.882131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86czq\" (UniqueName: \"kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq\") pod \"placementaa19-account-delete-twnzm\" (UID: \"943c9966-87bc-44fb-a19f-4ae15db1bdc4\") " pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.890330 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.911785 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.912045 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-kx2cq" podUID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" containerName="openstack-network-exporter" containerID="cri-o://ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5" gracePeriod=30 Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.941571 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.952145 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.958996 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jnfg\" (UniqueName: \"kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg\") pod \"glance6565-account-delete-q7tmg\" (UID: \"1da07093-5ea1-464a-a806-d30494ae9929\") " pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.961583 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.973089 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:25 crc kubenswrapper[4959]: I1003 13:56:25.988296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jnfg\" (UniqueName: \"kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg\") pod \"glance6565-account-delete-q7tmg\" (UID: \"1da07093-5ea1-464a-a806-d30494ae9929\") " pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.001967 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.007753 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="ovsdbserver-nb" containerID="cri-o://363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f" gracePeriod=300 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.065040 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qnfqp"] Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.065556 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.065771 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data podName:1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d nodeName:}" failed. No retries permitted until 2025-10-03 13:56:26.56575224 +0000 UTC m=+1555.769095657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data") pod "rabbitmq-cell1-server-0" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d") : configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.089160 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckqp\" (UniqueName: \"kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp\") pod \"cindercca9-account-delete-778b2\" (UID: \"f870f35a-1174-4568-9abb-b63d381483a2\") " pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.091150 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.091204 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data podName:d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2 nodeName:}" failed. No retries permitted until 2025-10-03 13:56:27.091190561 +0000 UTC m=+1556.294533978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data") pod "rabbitmq-server-0" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2") : configmap "rabbitmq-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.101387 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.127934 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qnfqp"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.133108 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.167283 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.193848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckqp\" (UniqueName: \"kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp\") pod \"cindercca9-account-delete-778b2\" (UID: \"f870f35a-1174-4568-9abb-b63d381483a2\") " pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.207251 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.208436 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.229720 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.238563 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="ovsdbserver-sb" containerID="cri-o://97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" gracePeriod=300 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.242704 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.244384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.245371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckqp\" (UniqueName: \"kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp\") pod \"cindercca9-account-delete-778b2\" (UID: \"f870f35a-1174-4568-9abb-b63d381483a2\") " pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.273076 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.289948 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.290162 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" containerID="cri-o://7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.290426 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="openstack-network-exporter" containerID="cri-o://9a443e77a12743b8133a947d16386d9a7160da93b531ed00b22855dabcff87e6" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.296324 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgwq7\" (UniqueName: \"kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7\") pod \"barbican2a29-account-delete-fhcv4\" (UID: \"53ba77ff-6205-4215-bb28-a45ea0b7b39f\") " pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.296393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dngf\" (UniqueName: \"kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf\") pod \"neutron3e00-account-delete-qthlw\" (UID: \"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b\") " pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.300282 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-cm7bx"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.316873 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-cm7bx"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.339259 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.340404 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.350421 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bfc789_a017_43e7_ba4f_96b93514bcf4.slice/crio-ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae7bb6a4_7314_482c_a09a_cb25b68d7e0a.slice/crio-conmon-363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae7bb6a4_7314_482c_a09a_cb25b68d7e0a.slice/crio-363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dbba504_e66c_4dbe_abe3_4e69ceec1a23.slice/crio-97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bfc789_a017_43e7_ba4f_96b93514bcf4.slice/crio-conmon-ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.372931 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.398498 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dngf\" (UniqueName: \"kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf\") pod \"neutron3e00-account-delete-qthlw\" (UID: \"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b\") " pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.398739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgwq7\" (UniqueName: \"kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7\") pod \"barbican2a29-account-delete-fhcv4\" (UID: \"53ba77ff-6205-4215-bb28-a45ea0b7b39f\") " pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.399336 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.399391 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:27.399374489 +0000 UTC m=+1556.602717906 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.405417 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-d78tr"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.435649 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.441823 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-d78tr"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.458265 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.459523 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.466109 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgwq7\" (UniqueName: \"kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7\") pod \"barbican2a29-account-delete-fhcv4\" (UID: \"53ba77ff-6205-4215-bb28-a45ea0b7b39f\") " pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.466178 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dngf\" (UniqueName: \"kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf\") pod \"neutron3e00-account-delete-qthlw\" (UID: \"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b\") " pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.469739 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-tdxpk"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.484336 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-tdxpk"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.502351 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.505852 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x8qd\" (UniqueName: \"kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd\") pod \"novacell0ef0b-account-delete-qnmr7\" (UID: \"3e788618-6d73-4794-8b04-ef41f70a94ab\") " pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.505943 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk544\" (UniqueName: \"kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544\") pod \"novaapi3cd4-account-delete-f4dn6\" (UID: \"a576d69d-a298-42d5-9c41-de82172c7a32\") " pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.522398 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-92xpj"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.536455 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-92xpj"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.588677 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.589183 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.611141 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk544\" (UniqueName: \"kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544\") pod \"novaapi3cd4-account-delete-f4dn6\" (UID: \"a576d69d-a298-42d5-9c41-de82172c7a32\") " pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.611495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x8qd\" (UniqueName: \"kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd\") pod \"novacell0ef0b-account-delete-qnmr7\" (UID: \"3e788618-6d73-4794-8b04-ef41f70a94ab\") " pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.615148 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: E1003 13:56:26.615201 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data podName:1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d nodeName:}" failed. No retries permitted until 2025-10-03 13:56:27.615184593 +0000 UTC m=+1556.818528010 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data") pod "rabbitmq-cell1-server-0" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d") : configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.637287 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.637769 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-server" containerID="cri-o://0b5395c8c9b277726906ea76a40f04683a601fd733480edda15aba4179fba2b5" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.638139 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="swift-recon-cron" containerID="cri-o://8d597ede10216948eca70fee9a0011ee11ae068e5bb066f3921145f9f96049e6" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.638188 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="rsync" containerID="cri-o://004034aa9b447a038bca1e3cffcf1b4fa3f5c22317a682340c9b6936b0225942" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.638243 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-expirer" containerID="cri-o://e93efd61469a9f4593c4dda8ba711b8d800d8760fc00d01dc03296a37a56265e" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.653576 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-auditor" containerID="cri-o://27d21e8e01afbe1d304e1134184443c9d6a5ca00c5cf331aa5561d1989f0cb19" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654007 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-updater" containerID="cri-o://30ca3bfd6f09224ce61168200372b479678d8278ab9c2dd7cba0dd99816f2b87" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654141 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-auditor" containerID="cri-o://a3d3d328227c680fc55d1e2c7cd597cae86b320ddc73849d22692bc5ba674e45" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654284 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-replicator" containerID="cri-o://6000516424e24bea6505665e93fef7a122ed78ffe7c2ba42d237e867092bc2dd" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654394 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-server" containerID="cri-o://23005919d7ab8893cd7eced93743eb0b32520fcd64407c9e2193608b3091452f" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654487 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-updater" containerID="cri-o://326cf0aa34a60cab6dd4c8c253341c34b40b93848f039d7eab65897e840f352f" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654612 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-reaper" containerID="cri-o://bed191a2d6ad3263372ef130bd3b54ebdd2a1a844aa9131fc1eac84370b2993e" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654731 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-replicator" containerID="cri-o://51c570be58868ca7518ace8ff746ead4f09d193a7fef31832eeba7dc57147fd9" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.654857 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-server" containerID="cri-o://e40663af33c9b4d1c19be468f7912534915bb3fba53c5989f5776aa52e428838" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.655001 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-replicator" containerID="cri-o://fc9e2bfb000766ae8566eb0974e5a4ea4297a6971e4c71891ab0065f78e71574" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.655123 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-auditor" containerID="cri-o://34ae11a48814fa28d689447bfba27eedc369d31d57a2d21cecd65f5c3ab9be8a" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.656346 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x8qd\" (UniqueName: \"kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd\") pod \"novacell0ef0b-account-delete-qnmr7\" (UID: \"3e788618-6d73-4794-8b04-ef41f70a94ab\") " pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.659625 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-s9dfg"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.660804 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk544\" (UniqueName: \"kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544\") pod \"novaapi3cd4-account-delete-f4dn6\" (UID: \"a576d69d-a298-42d5-9c41-de82172c7a32\") " pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.668590 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-s9dfg"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.739855 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.740085 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-86678496f6-dwd5r" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-log" containerID="cri-o://e2a2488abba43c5130cc5e66793ff02a224a83cad1a37062cb627eb282992260" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.740576 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-86678496f6-dwd5r" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-api" containerID="cri-o://617f60df069f8b127236c30ef29dbcf5ad86abf5cef1969b4f6275e6d0b31ca5" gracePeriod=30 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.750365 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-55hsl"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.768349 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-wr77d"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.785482 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kx2cq_a4bfc789-a017-43e7-ba4f-96b93514bcf4/openstack-network-exporter/0.log" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.785535 4959 generic.go:334] "Generic (PLEG): container finished" podID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" containerID="ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5" exitCode=2 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.785586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kx2cq" event={"ID":"a4bfc789-a017-43e7-ba4f-96b93514bcf4","Type":"ContainerDied","Data":"ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5"} Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.796452 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-55hsl"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.809042 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-wr77d"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.823322 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ae7bb6a4-7314-482c-a09a-cb25b68d7e0a/ovsdbserver-nb/0.log" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.823366 4959 generic.go:334] "Generic (PLEG): container finished" podID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerID="eba109515bfe8301850c53df177bc5231193d427a4c857c848ed3c8d12e4e21b" exitCode=2 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.823380 4959 generic.go:334] "Generic (PLEG): container finished" podID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerID="363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f" exitCode=143 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.824911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerDied","Data":"eba109515bfe8301850c53df177bc5231193d427a4c857c848ed3c8d12e4e21b"} Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.824937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerDied","Data":"363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f"} Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.921917 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.940137 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.971035 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.976274 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5dbba504-e66c-4dbe-abe3-4e69ceec1a23/ovsdbserver-sb/0.log" Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.976314 4959 generic.go:334] "Generic (PLEG): container finished" podID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerID="18ea0e8b0cc153504a9c772e75f38038e8891d7353b017f050ebb59b0ecbf766" exitCode=2 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.976332 4959 generic.go:334] "Generic (PLEG): container finished" podID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerID="97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" exitCode=143 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.976404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerDied","Data":"18ea0e8b0cc153504a9c772e75f38038e8891d7353b017f050ebb59b0ecbf766"} Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.976429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerDied","Data":"97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee"} Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.978329 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="dnsmasq-dns" containerID="cri-o://b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce" gracePeriod=10 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.982494 4959 generic.go:334] "Generic (PLEG): container finished" podID="ade995a8-1a92-43c8-9018-583d339f5f01" containerID="9a443e77a12743b8133a947d16386d9a7160da93b531ed00b22855dabcff87e6" exitCode=2 Oct 03 13:56:26 crc kubenswrapper[4959]: I1003 13:56:26.982537 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerDied","Data":"9a443e77a12743b8133a947d16386d9a7160da93b531ed00b22855dabcff87e6"} Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.043280 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-949rt"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.051508 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-949rt"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.064332 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.073970 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-aa19-account-create-j7rsg"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.081686 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-aa19-account-create-j7rsg"] Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.134302 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.134363 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data podName:d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2 nodeName:}" failed. No retries permitted until 2025-10-03 13:56:29.134348747 +0000 UTC m=+1558.337692164 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data") pod "rabbitmq-server-0" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2") : configmap "rabbitmq-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.135440 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.137851 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-log" containerID="cri-o://f2500fe9d0158d0ab935afc5fe14db033c74992aa7ae26ecba958ae3596c57f9" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.138292 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-httpd" containerID="cri-o://fc9d972ac9b2f06f0e0356aa259d27ac5d465cfc53bdfcc611986b0718292921" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.182922 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.183153 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="cinder-scheduler" containerID="cri-o://b304c4a31b60f0e9c502c5bfcdf56c0274ea52bcd6515a31c6e5973c7f3cb733" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.183287 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="probe" containerID="cri-o://0f239875c91ee1dd657942f1b4a5bee065b5388e0803e40d082ef09005a99e55" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.194222 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" containerID="cri-o://ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" gracePeriod=29 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.207604 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.210546 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76dd7cdf5-g59pr" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-api" containerID="cri-o://fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.210958 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76dd7cdf5-g59pr" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-httpd" containerID="cri-o://dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.227447 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.227642 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api-log" containerID="cri-o://cc4b4c055039a16bb8698b1f062f4ab7a7185cd1b62caa23f240835b1d1163dd" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.227970 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api" containerID="cri-o://a7c1d4fb80394150b035616306b0249611184085fee535c8c9ed09e63d2e3572" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.252160 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.253184 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-log" containerID="cri-o://a7b0ceeff61941890b21a6f7d2ca35938e4cdfd15247925f0209a5119fb4f1a1" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.255322 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-httpd" containerID="cri-o://9bc9b789b1e410f14438dd275ab4df1b4886b1a5b73a1ba9e3ec567e82067001" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.267317 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.277851 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6nt5n"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.297690 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6nt5n"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.315259 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cca9-account-create-tp77d"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.321468 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cca9-account-create-tp77d"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.328741 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.337274 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.337445 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerName="nova-scheduler-scheduler" containerID="cri-o://1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.354688 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-dxntm"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.369761 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kx2cq_a4bfc789-a017-43e7-ba4f-96b93514bcf4/openstack-network-exporter/0.log" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.369844 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.374276 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-dxntm"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.393971 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.423977 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="rabbitmq" containerID="cri-o://73f4e64df58083370f19d3a84fd7e6d994163b876b512f98d2ee1e62e5754b19" gracePeriod=604800 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.442812 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.442908 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.442942 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.442988 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.443039 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw968\" (UniqueName: \"kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.443062 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir\") pod \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\" (UID: \"a4bfc789-a017-43e7-ba4f-96b93514bcf4\") " Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.443981 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.444057 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:29.444040792 +0000 UTC m=+1558.647384209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.452336 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.452430 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.456861 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config" (OuterVolumeSpecName: "config") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.456925 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3e00-account-create-ps5fw"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.465998 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968" (OuterVolumeSpecName: "kube-api-access-pw968") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "kube-api-access-pw968". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.476757 4959 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 03 13:56:27 crc kubenswrapper[4959]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 03 13:56:27 crc kubenswrapper[4959]: + source /usr/local/bin/container-scripts/functions Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNBridge=br-int Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNRemote=tcp:localhost:6642 Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNEncapType=geneve Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNAvailabilityZones= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ EnableChassisAsGateway=true Oct 03 13:56:27 crc kubenswrapper[4959]: ++ PhysicalNetworks= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNHostName= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 03 13:56:27 crc kubenswrapper[4959]: ++ ovs_dir=/var/lib/openvswitch Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 03 13:56:27 crc kubenswrapper[4959]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + cleanup_ovsdb_server_semaphore Oct 03 13:56:27 crc kubenswrapper[4959]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 03 13:56:27 crc kubenswrapper[4959]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-2bhp5" message=< Oct 03 13:56:27 crc kubenswrapper[4959]: Exiting ovsdb-server (5) [ OK ] Oct 03 13:56:27 crc kubenswrapper[4959]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 03 13:56:27 crc kubenswrapper[4959]: + source /usr/local/bin/container-scripts/functions Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNBridge=br-int Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNRemote=tcp:localhost:6642 Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNEncapType=geneve Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNAvailabilityZones= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ EnableChassisAsGateway=true Oct 03 13:56:27 crc kubenswrapper[4959]: ++ PhysicalNetworks= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNHostName= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 03 13:56:27 crc kubenswrapper[4959]: ++ ovs_dir=/var/lib/openvswitch Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 03 13:56:27 crc kubenswrapper[4959]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + cleanup_ovsdb_server_semaphore Oct 03 13:56:27 crc kubenswrapper[4959]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 03 13:56:27 crc kubenswrapper[4959]: > Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.476802 4959 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 03 13:56:27 crc kubenswrapper[4959]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 03 13:56:27 crc kubenswrapper[4959]: + source /usr/local/bin/container-scripts/functions Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNBridge=br-int Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNRemote=tcp:localhost:6642 Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNEncapType=geneve Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNAvailabilityZones= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ EnableChassisAsGateway=true Oct 03 13:56:27 crc kubenswrapper[4959]: ++ PhysicalNetworks= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ OVNHostName= Oct 03 13:56:27 crc kubenswrapper[4959]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 03 13:56:27 crc kubenswrapper[4959]: ++ ovs_dir=/var/lib/openvswitch Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 03 13:56:27 crc kubenswrapper[4959]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 03 13:56:27 crc kubenswrapper[4959]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + sleep 0.5 Oct 03 13:56:27 crc kubenswrapper[4959]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 03 13:56:27 crc kubenswrapper[4959]: + cleanup_ovsdb_server_semaphore Oct 03 13:56:27 crc kubenswrapper[4959]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 03 13:56:27 crc kubenswrapper[4959]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 03 13:56:27 crc kubenswrapper[4959]: > pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" containerID="cri-o://68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.476872 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" containerID="cri-o://68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" gracePeriod=29 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.503319 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3e00-account-create-ps5fw"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.529769 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.538109 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.546893 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.546928 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.546942 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfc789-a017-43e7-ba4f-96b93514bcf4-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.546953 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw968\" (UniqueName: \"kubernetes.io/projected/a4bfc789-a017-43e7-ba4f-96b93514bcf4-kube-api-access-pw968\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.546966 4959 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a4bfc789-a017-43e7-ba4f-96b93514bcf4-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.556265 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.556530 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-log" containerID="cri-o://9f1517df64486c3bcc421232da8688f5478dc78c9cfeec76b3735eadbe8fbb25" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.556955 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-api" containerID="cri-o://72a3e7929b0adc554a8bfd78df9f7708f0538f33a7bbf4518d4a306cb7e072a8" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.573517 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8c4f-account-create-mhwmc"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.591346 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b4php"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.598541 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a4bfc789-a017-43e7-ba4f-96b93514bcf4" (UID: "a4bfc789-a017-43e7-ba4f-96b93514bcf4"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.644826 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee is running failed: container process not found" containerID="97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.645034 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee is running failed: container process not found" containerID="97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.645095 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8c4f-account-create-mhwmc"] Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.645576 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee is running failed: container process not found" containerID="97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.645600 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="ovsdbserver-sb" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.649518 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfc789-a017-43e7-ba4f-96b93514bcf4-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.649590 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: E1003 13:56:27.649632 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data podName:1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d nodeName:}" failed. No retries permitted until 2025-10-03 13:56:29.649619006 +0000 UTC m=+1558.852962423 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data") pod "rabbitmq-cell1-server-0" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d") : configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.666479 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b4php"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.681756 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.681950 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" containerID="cri-o://bd8aaedeb008acc8fdb85e7ff1a53c8c9fefc46446b823e52046fd7ccab71695" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.687042 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" containerID="cri-o://cfb72705ebccca6e3af246841cf201b6036478e1d451c879eb40ab8cf65d5799" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.748141 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="galera" containerID="cri-o://1af66b255906c8775c33d1bfe9a4fbf4e95006a4472a5b2389c20102dfd422b5" gracePeriod=30 Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.805442 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1433da4c-d3b1-48ea-b62e-e70af51671e5" path="/var/lib/kubelet/pods/1433da4c-d3b1-48ea-b62e-e70af51671e5/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.806082 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ec93e9-4e68-4d03-9ded-91a642ea9156" path="/var/lib/kubelet/pods/20ec93e9-4e68-4d03-9ded-91a642ea9156/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.806558 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8b2056-b39e-4134-84b4-c911d90a4958" path="/var/lib/kubelet/pods/2d8b2056-b39e-4134-84b4-c911d90a4958/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.807012 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0ebab1-59f6-4904-b830-b252e06f03e9" path="/var/lib/kubelet/pods/2f0ebab1-59f6-4904-b830-b252e06f03e9/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.807930 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="440fa439-935e-4cba-8da7-89b2f6a797ea" path="/var/lib/kubelet/pods/440fa439-935e-4cba-8da7-89b2f6a797ea/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.808815 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b37f5d-cfd9-4bad-a768-f9c2d40f279c" path="/var/lib/kubelet/pods/51b37f5d-cfd9-4bad-a768-f9c2d40f279c/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.811720 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56a3ee2c-e196-4d8d-a796-507963138289" path="/var/lib/kubelet/pods/56a3ee2c-e196-4d8d-a796-507963138289/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.812222 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58529f64-d1c9-4feb-ba01-a0cdea44996e" path="/var/lib/kubelet/pods/58529f64-d1c9-4feb-ba01-a0cdea44996e/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.812668 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58ba97c9-9be2-41bd-a165-e3baecf9c97a" path="/var/lib/kubelet/pods/58ba97c9-9be2-41bd-a165-e3baecf9c97a/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.813099 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae02254-e00c-447a-84a9-4ab1b433b5e8" path="/var/lib/kubelet/pods/6ae02254-e00c-447a-84a9-4ab1b433b5e8/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.817360 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e634f23-0fe8-438b-9657-d719b3f01205" path="/var/lib/kubelet/pods/6e634f23-0fe8-438b-9657-d719b3f01205/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.818227 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9655f477-fe02-412d-b52d-6aea687e1e57" path="/var/lib/kubelet/pods/9655f477-fe02-412d-b52d-6aea687e1e57/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.818518 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5dbba504-e66c-4dbe-abe3-4e69ceec1a23/ovsdbserver-sb/0.log" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.818579 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.828701 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19f4ff5-e607-4a03-92c0-6365e6d93bf6" path="/var/lib/kubelet/pods/c19f4ff5-e607-4a03-92c0-6365e6d93bf6/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.830260 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c41cfce6-87b2-4ff2-8697-e739f3fa1ed7" path="/var/lib/kubelet/pods/c41cfce6-87b2-4ff2-8697-e739f3fa1ed7/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.830888 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b4233a-0bb4-41bc-a81f-1a299aa9308d" path="/var/lib/kubelet/pods/e2b4233a-0bb4-41bc-a81f-1a299aa9308d/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.831374 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2448b40-93ae-4b63-b494-fe294dfcdb35" path="/var/lib/kubelet/pods/f2448b40-93ae-4b63-b494-fe294dfcdb35/volumes" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.832612 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-mksg4"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.832639 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-mksg4"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.832653 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2a29-account-create-vr7dm"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.832663 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2a29-account-create-vr7dm"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.832672 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-t4tw9"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.853532 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.919937 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-t4tw9"] Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.942064 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993444 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993490 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zmsd\" (UniqueName: \"kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993657 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993693 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993711 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993765 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.993816 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config\") pod \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\" (UID: \"5dbba504-e66c-4dbe-abe3-4e69ceec1a23\") " Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.994825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config" (OuterVolumeSpecName: "config") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:27 crc kubenswrapper[4959]: I1003 13:56:27.995472 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts" (OuterVolumeSpecName: "scripts") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.002840 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3cd4-account-create-6tgxr"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.003264 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.012337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd" (OuterVolumeSpecName: "kube-api-access-4zmsd") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "kube-api-access-4zmsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.012438 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.018755 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3cd4-account-create-6tgxr"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.020334 4959 generic.go:334] "Generic (PLEG): container finished" podID="a7088287-7362-42f5-987c-be9f700c50cb" containerID="bae04dd3c4a3a80e408d2aa1da0eb1a0f82c686e02752a0f01f6986ff0f0542b" exitCode=137 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.020489 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.022854 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d265899-c75a-462f-9779-c573bd35a8d4" containerID="bd8aaedeb008acc8fdb85e7ff1a53c8c9fefc46446b823e52046fd7ccab71695" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.022903 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerDied","Data":"bd8aaedeb008acc8fdb85e7ff1a53c8c9fefc46446b823e52046fd7ccab71695"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.028849 4959 generic.go:334] "Generic (PLEG): container finished" podID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerID="f2500fe9d0158d0ab935afc5fe14db033c74992aa7ae26ecba958ae3596c57f9" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.028930 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerDied","Data":"f2500fe9d0158d0ab935afc5fe14db033c74992aa7ae26ecba958ae3596c57f9"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.029552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-p2q4s"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.041060 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.041118 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-p2q4s"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.052379 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.063833 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ef0b-account-create-4xq88"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.071200 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ef0b-account-create-4xq88"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072597 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="004034aa9b447a038bca1e3cffcf1b4fa3f5c22317a682340c9b6936b0225942" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072630 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="e93efd61469a9f4593c4dda8ba711b8d800d8760fc00d01dc03296a37a56265e" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072643 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="30ca3bfd6f09224ce61168200372b479678d8278ab9c2dd7cba0dd99816f2b87" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072653 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="a3d3d328227c680fc55d1e2c7cd597cae86b320ddc73849d22692bc5ba674e45" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072661 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="6000516424e24bea6505665e93fef7a122ed78ffe7c2ba42d237e867092bc2dd" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072672 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="23005919d7ab8893cd7eced93743eb0b32520fcd64407c9e2193608b3091452f" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072680 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="326cf0aa34a60cab6dd4c8c253341c34b40b93848f039d7eab65897e840f352f" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072688 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="27d21e8e01afbe1d304e1134184443c9d6a5ca00c5cf331aa5561d1989f0cb19" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072696 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="51c570be58868ca7518ace8ff746ead4f09d193a7fef31832eeba7dc57147fd9" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072704 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="e40663af33c9b4d1c19be468f7912534915bb3fba53c5989f5776aa52e428838" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072712 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="bed191a2d6ad3263372ef130bd3b54ebdd2a1a844aa9131fc1eac84370b2993e" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072720 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="34ae11a48814fa28d689447bfba27eedc369d31d57a2d21cecd65f5c3ab9be8a" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072727 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="fc9e2bfb000766ae8566eb0974e5a4ea4297a6971e4c71891ab0065f78e71574" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072735 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="0b5395c8c9b277726906ea76a40f04683a601fd733480edda15aba4179fba2b5" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072815 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"004034aa9b447a038bca1e3cffcf1b4fa3f5c22317a682340c9b6936b0225942"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072846 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"e93efd61469a9f4593c4dda8ba711b8d800d8760fc00d01dc03296a37a56265e"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072860 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"30ca3bfd6f09224ce61168200372b479678d8278ab9c2dd7cba0dd99816f2b87"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072870 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"a3d3d328227c680fc55d1e2c7cd597cae86b320ddc73849d22692bc5ba674e45"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072881 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"6000516424e24bea6505665e93fef7a122ed78ffe7c2ba42d237e867092bc2dd"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072891 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"23005919d7ab8893cd7eced93743eb0b32520fcd64407c9e2193608b3091452f"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072902 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"326cf0aa34a60cab6dd4c8c253341c34b40b93848f039d7eab65897e840f352f"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072913 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"27d21e8e01afbe1d304e1134184443c9d6a5ca00c5cf331aa5561d1989f0cb19"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"51c570be58868ca7518ace8ff746ead4f09d193a7fef31832eeba7dc57147fd9"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072934 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"e40663af33c9b4d1c19be468f7912534915bb3fba53c5989f5776aa52e428838"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072944 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"bed191a2d6ad3263372ef130bd3b54ebdd2a1a844aa9131fc1eac84370b2993e"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072955 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"34ae11a48814fa28d689447bfba27eedc369d31d57a2d21cecd65f5c3ab9be8a"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072966 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"fc9e2bfb000766ae8566eb0974e5a4ea4297a6971e4c71891ab0065f78e71574"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.072976 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"0b5395c8c9b277726906ea76a40f04683a601fd733480edda15aba4179fba2b5"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.077386 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.079438 4959 generic.go:334] "Generic (PLEG): container finished" podID="303c0421-debb-437b-8a12-4ff587f419f3" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.079490 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerDied","Data":"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.086972 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.088235 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener-log" containerID="cri-o://023f6f41efd8b70bb5c4bc20ca5e840eaa4f8d065b77a0f728e8d0ecf836a688" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.088800 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener" containerID="cri-o://48413a96dbd55c4eac4e40c0f894a4fe721bcad741ce542a08321841cf77afd6" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.090653 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.091894 4959 generic.go:334] "Generic (PLEG): container finished" podID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerID="a7b0ceeff61941890b21a6f7d2ca35938e4cdfd15247925f0209a5119fb4f1a1" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.093048 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerDied","Data":"a7b0ceeff61941890b21a6f7d2ca35938e4cdfd15247925f0209a5119fb4f1a1"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.095899 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.095955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zbxp\" (UniqueName: \"kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096059 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096109 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096143 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096160 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb\") pod \"fb01d41b-854a-4072-8b05-807edc22812f\" (UID: \"fb01d41b-854a-4072-8b05-807edc22812f\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096605 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096622 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096639 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096650 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096658 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.096666 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zmsd\" (UniqueName: \"kubernetes.io/projected/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-kube-api-access-4zmsd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.108307 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.115821 4959 generic.go:334] "Generic (PLEG): container finished" podID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerID="cc4b4c055039a16bb8698b1f062f4ab7a7185cd1b62caa23f240835b1d1163dd" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.115904 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerDied","Data":"cc4b4c055039a16bb8698b1f062f4ab7a7185cd1b62caa23f240835b1d1163dd"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.124702 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.124956 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-54b8ffdcf-7hs45" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker-log" containerID="cri-o://d5033b93560574e923577b90dc0e6761b4304b82f8fedaf8d5ae2041e3f74a66" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.125464 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-54b8ffdcf-7hs45" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker" containerID="cri-o://95dbbc2b264de16726fd5cf962dcf5b2ea23cce70901b014633e800857d2e6ac" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.130883 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.136256 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.136511 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-677d776894-ljrrj" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api-log" containerID="cri-o://1f8f42d263cd44bb7f3d741af35c4900133542e621536ce527da52ab82dbee5c" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.136884 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-677d776894-ljrrj" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api" containerID="cri-o://06d4a6beaf8c7baed60c86f5ef66a97b9dcd29bdf45ef077eba2ec59e7cf24e6" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.146005 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.146226 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e9f6f6f4ded33babe3c6b4e9a7384c3513fe53da441505d10b5d13c79bd8152e" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.146623 4959 generic.go:334] "Generic (PLEG): container finished" podID="fb01d41b-854a-4072-8b05-807edc22812f" containerID="b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.147047 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.147531 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" event={"ID":"fb01d41b-854a-4072-8b05-807edc22812f","Type":"ContainerDied","Data":"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.147563 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-fssm2" event={"ID":"fb01d41b-854a-4072-8b05-807edc22812f","Type":"ContainerDied","Data":"4e490e880c7b9fe405ba3fd81d5dcffec829e669e020897e0d7009a2412279aa"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.147583 4959 scope.go:117] "RemoveContainer" containerID="b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.163526 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.171015 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.172356 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5dbba504-e66c-4dbe-abe3-4e69ceec1a23/ovsdbserver-sb/0.log" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.172422 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5dbba504-e66c-4dbe-abe3-4e69ceec1a23","Type":"ContainerDied","Data":"2594e4b02494d055c82a0fcaa360f604167c2d72ae3b70140a42cc2607d7e46c"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.172544 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.173889 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp" (OuterVolumeSpecName: "kube-api-access-5zbxp") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "kube-api-access-5zbxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.176133 4959 generic.go:334] "Generic (PLEG): container finished" podID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerID="dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6" exitCode=0 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.176202 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerDied","Data":"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.181295 4959 generic.go:334] "Generic (PLEG): container finished" podID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerID="e2a2488abba43c5130cc5e66793ff02a224a83cad1a37062cb627eb282992260" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.181425 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerDied","Data":"e2a2488abba43c5130cc5e66793ff02a224a83cad1a37062cb627eb282992260"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.182527 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerStarted","Data":"e1a6304ec8e0c99f5267c0f6e0fb012f9e604ea3847f5b4fdbea30d06ae23ecf"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.183383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerStarted","Data":"32ab7174359f57de9bc9ba592a08d7057265165d78c313d0b5ca8cbb0551bb15"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.185430 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lpc2k"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.192704 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.192899 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.193728 4959 generic.go:334] "Generic (PLEG): container finished" podID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerID="9f1517df64486c3bcc421232da8688f5478dc78c9cfeec76b3735eadbe8fbb25" exitCode=143 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.193796 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerDied","Data":"9f1517df64486c3bcc421232da8688f5478dc78c9cfeec76b3735eadbe8fbb25"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.198376 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zbxp\" (UniqueName: \"kubernetes.io/projected/fb01d41b-854a-4072-8b05-807edc22812f-kube-api-access-5zbxp\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.205193 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kx2cq_a4bfc789-a017-43e7-ba4f-96b93514bcf4/openstack-network-exporter/0.log" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.205259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kx2cq" event={"ID":"a4bfc789-a017-43e7-ba4f-96b93514bcf4","Type":"ContainerDied","Data":"bf8ac294982419fa3eb8f1a46d0985ec703ea43985cbf877285ce1bc71cb906b"} Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.205368 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kx2cq" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.207630 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lpc2k"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.217121 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.217515 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.228659 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6pgdx"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.233966 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6pgdx"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.260507 4959 scope.go:117] "RemoveContainer" containerID="8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.272018 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ae7bb6a4-7314-482c-a09a-cb25b68d7e0a/ovsdbserver-nb/0.log" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.272083 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.275319 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.285975 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.297738 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.347758 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.355026 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-kx2cq"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.395272 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="rabbitmq" containerID="cri-o://79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6" gracePeriod=604800 Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.395380 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.395343 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401469 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401554 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config\") pod \"a7088287-7362-42f5-987c-be9f700c50cb\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401624 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401761 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret\") pod \"a7088287-7362-42f5-987c-be9f700c50cb\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401798 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2j7v\" (UniqueName: \"kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401816 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401901 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config\") pod \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\" (UID: \"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5lw6\" (UniqueName: \"kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6\") pod \"a7088287-7362-42f5-987c-be9f700c50cb\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.401976 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle\") pod \"a7088287-7362-42f5-987c-be9f700c50cb\" (UID: \"a7088287-7362-42f5-987c-be9f700c50cb\") " Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.404420 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.404572 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.421717 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.427399 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.432299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts" (OuterVolumeSpecName: "scripts") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.442511 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config" (OuterVolumeSpecName: "config") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.459441 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.459494 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.459918 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.459940 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.472285 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.487270 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.488226 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v" (OuterVolumeSpecName: "kube-api-access-v2j7v") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "kube-api-access-v2j7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.488317 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.492168 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6" (OuterVolumeSpecName: "kube-api-access-x5lw6") pod "a7088287-7362-42f5-987c-be9f700c50cb" (UID: "a7088287-7362-42f5-987c-be9f700c50cb"). InnerVolumeSpecName "kube-api-access-x5lw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.505295 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.513963 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.515669 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.530167 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.530580 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5b84954875-j5bdp" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-httpd" containerID="cri-o://802921468f08beb7a7faecf994bffea767c35ceda75b15525c02a4e2c871e13a" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.530995 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531034 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531048 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2j7v\" (UniqueName: \"kubernetes.io/projected/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-kube-api-access-v2j7v\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531058 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531070 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531080 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5lw6\" (UniqueName: \"kubernetes.io/projected/a7088287-7362-42f5-987c-be9f700c50cb-kube-api-access-x5lw6\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.531541 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5b84954875-j5bdp" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-server" containerID="cri-o://e51713b9b874f135af735bd09481738fbf57ba1d4691649ef78b87b0fca43cd1" gracePeriod=30 Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.539110 4959 scope.go:117] "RemoveContainer" containerID="b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.539702 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce\": container with ID starting with b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce not found: ID does not exist" containerID="b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.539830 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce"} err="failed to get container status \"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce\": rpc error: code = NotFound desc = could not find container \"b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce\": container with ID starting with b08d124609f35028728ece887b455e3cfe46b182ec63bcc0366cef0a0236e3ce not found: ID does not exist" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.539931 4959 scope.go:117] "RemoveContainer" containerID="8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.540180 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf\": container with ID starting with 8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf not found: ID does not exist" containerID="8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.540309 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf"} err="failed to get container status \"8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf\": rpc error: code = NotFound desc = could not find container \"8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf\": container with ID starting with 8227c92128f896e7fb4f5a20c046dc8977e9e77221814eaebe922dbe883dbfcf not found: ID does not exist" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.540372 4959 scope.go:117] "RemoveContainer" containerID="18ea0e8b0cc153504a9c772e75f38038e8891d7353b017f050ebb59b0ecbf766" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.573695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config" (OuterVolumeSpecName: "config") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.612533 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.625632 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.625729 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.629775 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.630381 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.633519 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.652766 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.659002 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.711667 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.714150 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: W1003 13:56:28.721042 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e788618_6d73_4794_8b04_ef41f70a94ab.slice/crio-8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be WatchSource:0}: Error finding container 8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be: Status 404 returned error can't find the container with id 8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.721247 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:28 crc kubenswrapper[4959]: E1003 13:56:28.721282 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.753302 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.908916 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7088287-7362-42f5-987c-be9f700c50cb" (UID: "a7088287-7362-42f5-987c-be9f700c50cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.912016 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.921579 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.935685 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.950188 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "5dbba504-e66c-4dbe-abe3-4e69ceec1a23" (UID: "5dbba504-e66c-4dbe-abe3-4e69ceec1a23"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.956055 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5b84954875-j5bdp" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.157:8080/healthcheck\": dial tcp 10.217.0.157:8080: connect: connection refused" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.956416 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.956819 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5b84954875-j5bdp" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.157:8080/healthcheck\": dial tcp 10.217.0.157:8080: connect: connection refused" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957810 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957829 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dbba504-e66c-4dbe-abe3-4e69ceec1a23-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957838 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957847 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957857 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.957865 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.969357 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a7088287-7362-42f5-987c-be9f700c50cb" (UID: "a7088287-7362-42f5-987c-be9f700c50cb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:28 crc kubenswrapper[4959]: I1003 13:56:28.988291 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.049543 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fb01d41b-854a-4072-8b05-807edc22812f" (UID: "fb01d41b-854a-4072-8b05-807edc22812f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.060281 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.060714 4959 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.074417 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb01d41b-854a-4072-8b05-807edc22812f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.101093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.153920 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" (UID: "ae7bb6a4-7314-482c-a09a-cb25b68d7e0a"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.177262 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.177290 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.177347 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.177389 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data podName:d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2 nodeName:}" failed. No retries permitted until 2025-10-03 13:56:33.177375783 +0000 UTC m=+1562.380719200 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data") pod "rabbitmq-server-0" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2") : configmap "rabbitmq-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.182959 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a7088287-7362-42f5-987c-be9f700c50cb" (UID: "a7088287-7362-42f5-987c-be9f700c50cb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.237654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3cd4-account-delete-f4dn6" event={"ID":"a576d69d-a298-42d5-9c41-de82172c7a32","Type":"ContainerStarted","Data":"a29734fee59c9beeb03f3ad6d1144d2d4658d86e65b681058373ba43b2de9a9b"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.240651 4959 generic.go:334] "Generic (PLEG): container finished" podID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerID="d5033b93560574e923577b90dc0e6761b4304b82f8fedaf8d5ae2041e3f74a66" exitCode=143 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.240737 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerDied","Data":"d5033b93560574e923577b90dc0e6761b4304b82f8fedaf8d5ae2041e3f74a66"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.244225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6565-account-delete-q7tmg" event={"ID":"1da07093-5ea1-464a-a806-d30494ae9929","Type":"ContainerStarted","Data":"5d8bbcf7de2d7125a34188d0b847f27d498b39bdb6213d69b5bd4d7bd26850c4"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.246574 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerStarted","Data":"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.248062 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerStarted","Data":"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.250086 4959 generic.go:334] "Generic (PLEG): container finished" podID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerID="1af66b255906c8775c33d1bfe9a4fbf4e95006a4472a5b2389c20102dfd422b5" exitCode=0 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.250149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerDied","Data":"1af66b255906c8775c33d1bfe9a4fbf4e95006a4472a5b2389c20102dfd422b5"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.250172 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"db9f7e3a-3cea-4bfd-986b-3a026d04075f","Type":"ContainerDied","Data":"6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.250184 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6962cca4e62c4da527581576d22bda7327b7e54b01d287071f619eb3179dc3de" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.256032 4959 generic.go:334] "Generic (PLEG): container finished" podID="2ec784f6-4b46-435e-af12-e899033f8348" containerID="023f6f41efd8b70bb5c4bc20ca5e840eaa4f8d065b77a0f728e8d0ecf836a688" exitCode=143 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.256414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerDied","Data":"023f6f41efd8b70bb5c4bc20ca5e840eaa4f8d065b77a0f728e8d0ecf836a688"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.259241 4959 generic.go:334] "Generic (PLEG): container finished" podID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerID="0f239875c91ee1dd657942f1b4a5bee065b5388e0803e40d082ef09005a99e55" exitCode=0 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.259319 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerDied","Data":"0f239875c91ee1dd657942f1b4a5bee065b5388e0803e40d082ef09005a99e55"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.262420 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ae7bb6a4-7314-482c-a09a-cb25b68d7e0a/ovsdbserver-nb/0.log" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.262632 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.262871 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae7bb6a4-7314-482c-a09a-cb25b68d7e0a","Type":"ContainerDied","Data":"4e339b80df0668dff924d38130d42e6972293d3ab0b65a87d13a033ae0d1d964"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.264245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementaa19-account-delete-twnzm" event={"ID":"943c9966-87bc-44fb-a19f-4ae15db1bdc4","Type":"ContainerStarted","Data":"255a0bea8e5d0b5422082e03fdb0393292916b72890c8510403dc958dc9ecd28"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.274840 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.276872 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercca9-account-delete-778b2" event={"ID":"f870f35a-1174-4568-9abb-b63d381483a2","Type":"ContainerStarted","Data":"5c0c1f1ab4ccf63b22028b84710b27349347718809db875c9643fa860262fd5a"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.278424 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a7088287-7362-42f5-987c-be9f700c50cb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.279737 4959 generic.go:334] "Generic (PLEG): container finished" podID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerID="e51713b9b874f135af735bd09481738fbf57ba1d4691649ef78b87b0fca43cd1" exitCode=0 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.279765 4959 generic.go:334] "Generic (PLEG): container finished" podID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerID="802921468f08beb7a7faecf994bffea767c35ceda75b15525c02a4e2c871e13a" exitCode=0 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.279806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerDied","Data":"e51713b9b874f135af735bd09481738fbf57ba1d4691649ef78b87b0fca43cd1"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.279830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerDied","Data":"802921468f08beb7a7faecf994bffea767c35ceda75b15525c02a4e2c871e13a"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.281554 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef0b-account-delete-qnmr7" event={"ID":"3e788618-6d73-4794-8b04-ef41f70a94ab","Type":"ContainerStarted","Data":"8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.283033 4959 generic.go:334] "Generic (PLEG): container finished" podID="56e5d3a5-a733-402b-b393-d07706d30c48" containerID="e9f6f6f4ded33babe3c6b4e9a7384c3513fe53da441505d10b5d13c79bd8152e" exitCode=0 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.283138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"56e5d3a5-a733-402b-b393-d07706d30c48","Type":"ContainerDied","Data":"e9f6f6f4ded33babe3c6b4e9a7384c3513fe53da441505d10b5d13c79bd8152e"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.288666 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2a29-account-delete-fhcv4" event={"ID":"53ba77ff-6205-4215-bb28-a45ea0b7b39f","Type":"ContainerStarted","Data":"33e104eb4372325d59ebd0f1bc19b917f3a71beed66561e03bdbe61fe52d1595"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.295752 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerID="1f8f42d263cd44bb7f3d741af35c4900133542e621536ce527da52ab82dbee5c" exitCode=143 Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.295835 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerDied","Data":"1f8f42d263cd44bb7f3d741af35c4900133542e621536ce527da52ab82dbee5c"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.297671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3e00-account-delete-qthlw" event={"ID":"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b","Type":"ContainerStarted","Data":"11a81c2ca063c7d4aff8760ea97e39f85321948f51643fd76af45f40c8334c6f"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.298529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerStarted","Data":"7857bb4799b289a4167000cc02e9f5e9f0f6f89a66c11cc29705b6cc1a5b7f19"} Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.375839 4959 scope.go:117] "RemoveContainer" containerID="97c3eb75e9791688146ffbbfa93b54f17a22455ddcb09157d6e25464c0c458ee" Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.489883 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.490232 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:33.490217934 +0000 UTC m=+1562.693561351 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.513689 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.655638 4959 scope.go:117] "RemoveContainer" containerID="ddc3afa5ee698a4c410494d68d5db95174eaea0b7d41c91e6601c4ca40121ee5" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.691920 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.692971 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhljg\" (UniqueName: \"kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693092 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693130 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693227 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693270 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693306 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.693335 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts\") pod \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\" (UID: \"db9f7e3a-3cea-4bfd-986b-3a026d04075f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.696178 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.700166 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.700975 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.701068 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: E1003 13:56:29.701130 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data podName:1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d nodeName:}" failed. No retries permitted until 2025-10-03 13:56:33.701098818 +0000 UTC m=+1562.904442235 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data") pod "rabbitmq-cell1-server-0" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d") : configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.709499 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets" (OuterVolumeSpecName: "secrets") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.710476 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5960139e-48e3-4a67-981f-539d78fb46a1" path="/var/lib/kubelet/pods/5960139e-48e3-4a67-981f-539d78fb46a1/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.711732 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7830ad44-ea59-47a8-973b-4438930500dc" path="/var/lib/kubelet/pods/7830ad44-ea59-47a8-973b-4438930500dc/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.712307 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" path="/var/lib/kubelet/pods/a4bfc789-a017-43e7-ba4f-96b93514bcf4/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.713285 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7088287-7362-42f5-987c-be9f700c50cb" path="/var/lib/kubelet/pods/a7088287-7362-42f5-987c-be9f700c50cb/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.713793 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2f39eb9-4894-4b1c-9676-c0e4485932c5" path="/var/lib/kubelet/pods/c2f39eb9-4894-4b1c-9676-c0e4485932c5/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.714421 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.714476 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc575e3-fc82-48e4-ae54-27f374e37063" path="/var/lib/kubelet/pods/cdc575e3-fc82-48e4-ae54-27f374e37063/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.715662 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceeec8b6-407f-4a35-a745-d51f676bcbaf" path="/var/lib/kubelet/pods/ceeec8b6-407f-4a35-a745-d51f676bcbaf/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.716824 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg" (OuterVolumeSpecName: "kube-api-access-jhljg") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "kube-api-access-jhljg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.717080 4959 scope.go:117] "RemoveContainer" containerID="eba109515bfe8301850c53df177bc5231193d427a4c857c848ed3c8d12e4e21b" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.717231 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c3386b-7ed5-4876-bbcb-f00137430c8b" path="/var/lib/kubelet/pods/d4c3386b-7ed5-4876-bbcb-f00137430c8b/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.718751 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.719123 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d93957ba-a96d-4a52-ab5c-46c80c04f451" path="/var/lib/kubelet/pods/d93957ba-a96d-4a52-ab5c-46c80c04f451/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.720742 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbc3e407-060d-4f10-8993-71f733e510ed" path="/var/lib/kubelet/pods/dbc3e407-060d-4f10-8993-71f733e510ed/volumes" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.754391 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.780347 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795362 4959 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795392 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795402 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795421 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795431 4959 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db9f7e3a-3cea-4bfd-986b-3a026d04075f-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795439 4959 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795448 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhljg\" (UniqueName: \"kubernetes.io/projected/db9f7e3a-3cea-4bfd-986b-3a026d04075f-kube-api-access-jhljg\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.795474 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.822111 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.824994 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.833458 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "db9f7e3a-3cea-4bfd-986b-3a026d04075f" (UID: "db9f7e3a-3cea-4bfd-986b-3a026d04075f"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.854680 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897417 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897481 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs\") pod \"56e5d3a5-a733-402b-b393-d07706d30c48\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897578 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897676 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897726 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data\") pod \"56e5d3a5-a733-402b-b393-d07706d30c48\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897757 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs\") pod \"56e5d3a5-a733-402b-b393-d07706d30c48\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897781 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw9sb\" (UniqueName: \"kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb\") pod \"56e5d3a5-a733-402b-b393-d07706d30c48\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897809 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897839 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897868 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897889 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km9l4\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897915 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd\") pod \"f325dd29-3b0d-4188-9ac2-e681d56e881f\" (UID: \"f325dd29-3b0d-4188-9ac2-e681d56e881f\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897932 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle\") pod \"56e5d3a5-a733-402b-b393-d07706d30c48\" (UID: \"56e5d3a5-a733-402b-b393-d07706d30c48\") " Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.897954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.898366 4959 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/db9f7e3a-3cea-4bfd-986b-3a026d04075f-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.898382 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.898401 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.902510 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-86678496f6-dwd5r" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.141:8778/\": read tcp 10.217.0.2:58556->10.217.0.141:8778: read: connection reset by peer" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.902815 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-86678496f6-dwd5r" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.141:8778/\": read tcp 10.217.0.2:58542->10.217.0.141:8778: read: connection reset by peer" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.904556 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.910327 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4" (OuterVolumeSpecName: "kube-api-access-km9l4") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "kube-api-access-km9l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.928181 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb" (OuterVolumeSpecName: "kube-api-access-tw9sb") pod "56e5d3a5-a733-402b-b393-d07706d30c48" (UID: "56e5d3a5-a733-402b-b393-d07706d30c48"). InnerVolumeSpecName "kube-api-access-tw9sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.931762 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.963437 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data" (OuterVolumeSpecName: "config-data") pod "56e5d3a5-a733-402b-b393-d07706d30c48" (UID: "56e5d3a5-a733-402b-b393-d07706d30c48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:29 crc kubenswrapper[4959]: I1003 13:56:29.996354 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data" (OuterVolumeSpecName: "config-data") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000576 4959 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000602 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000618 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw9sb\" (UniqueName: \"kubernetes.io/projected/56e5d3a5-a733-402b-b393-d07706d30c48-kube-api-access-tw9sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000632 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000643 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km9l4\" (UniqueName: \"kubernetes.io/projected/f325dd29-3b0d-4188-9ac2-e681d56e881f-kube-api-access-km9l4\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.000655 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f325dd29-3b0d-4188-9ac2-e681d56e881f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.004337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "56e5d3a5-a733-402b-b393-d07706d30c48" (UID: "56e5d3a5-a733-402b-b393-d07706d30c48"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.024292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56e5d3a5-a733-402b-b393-d07706d30c48" (UID: "56e5d3a5-a733-402b-b393-d07706d30c48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.026955 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.032474 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "56e5d3a5-a733-402b-b393-d07706d30c48" (UID: "56e5d3a5-a733-402b-b393-d07706d30c48"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.046887 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.053380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f325dd29-3b0d-4188-9ac2-e681d56e881f" (UID: "f325dd29-3b0d-4188-9ac2-e681d56e881f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108748 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108827 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108852 4959 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108862 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108871 4959 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/56e5d3a5-a733-402b-b393-d07706d30c48-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.108881 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f325dd29-3b0d-4188-9ac2-e681d56e881f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131600 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131646 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-fssm2"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131668 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131681 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131694 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131708 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.131730 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.132046 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-central-agent" containerID="cri-o://4ddb932943c117d08c1006b64f264db30d2dae0973fb4d5017d89ad84b3ec3a8" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.132562 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="proxy-httpd" containerID="cri-o://ada15b2c15fe2f2dacc57422fd81cb3d234b56614971b21c30ab71da932c7b32" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.132621 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="sg-core" containerID="cri-o://13ffbebcbaac00743c439cf6bbe0144cf9261560d362f82d27b626d91213b204" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.132668 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-notification-agent" containerID="cri-o://47160c83891466f0f38183ffc99e4cb435c5e8b448cf8d1325e15334ebeb44d0" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.161572 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.161761 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="593264f5-6188-4861-9d83-6918034eb46e" containerName="kube-state-metrics" containerID="cri-o://437ecf2e1e223998fb6b040d3dc342771553ad152a649a307ba005f342deed15" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.203456 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.237173 4959 scope.go:117] "RemoveContainer" containerID="363fe17448132174ab0a390380abf4f5f411e67d01b8a21ce7dc914acc3b648f" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.252316 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.252549 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="30542159-188f-4508-a1b4-70285f59aef3" containerName="memcached" containerID="cri-o://aae8036f06d6bb33643dcc5d309e687768c23fb0437205d1473dc1f70e1e4751" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.329270 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-spfmm"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.352782 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xbd64"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.364980 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xbd64"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.372399 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-spfmm"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.380015 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.381089 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5b6c4d45db-cwzjn" podUID="f51595d7-80bf-4436-9429-9f0da918850d" containerName="keystone-api" containerID="cri-o://964de272e99fd5d3325f928f234440a05a23315ec4d4c92db1d5a71e50facdcf" gracePeriod=30 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.393716 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.429201 4959 generic.go:334] "Generic (PLEG): container finished" podID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerID="617f60df069f8b127236c30ef29dbcf5ad86abf5cef1969b4f6275e6d0b31ca5" exitCode=0 Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.429311 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerDied","Data":"617f60df069f8b127236c30ef29dbcf5ad86abf5cef1969b4f6275e6d0b31ca5"} Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.445714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b84954875-j5bdp" event={"ID":"f325dd29-3b0d-4188-9ac2-e681d56e881f","Type":"ContainerDied","Data":"d0cfb94ed897d811d8aa2d0ef9731b2fa0437d894fa179674d671b3d8965b0cc"} Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.445989 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b84954875-j5bdp" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.462541 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2pszr"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.469798 4959 scope.go:117] "RemoveContainer" containerID="bae04dd3c4a3a80e408d2aa1da0eb1a0f82c686e02752a0f01f6986ff0f0542b" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.491380 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.491707 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.493263 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"56e5d3a5-a733-402b-b393-d07706d30c48","Type":"ContainerDied","Data":"cb13e6ee4cff2a3395b564828dea636a2d16c60883906eb1fff7b014b19f128d"} Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.500933 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2pszr"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.516361 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7408-account-create-5dpjp"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.528539 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.168:9292/healthcheck\": dial tcp 10.217.0.168:9292: connect: connection refused" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.528812 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.168:9292/healthcheck\": dial tcp 10.217.0.168:9292: connect: connection refused" Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.529744 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7408-account-create-5dpjp"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.537376 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.546521 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-5b84954875-j5bdp"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.586038 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2jgb6"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.590597 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2jgb6"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.596414 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6565-account-create-8k2w6"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.600639 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:30 crc kubenswrapper[4959]: I1003 13:56:30.609833 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6565-account-create-8k2w6"] Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.141946 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" containerID="cri-o://f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" gracePeriod=30 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.184377 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:37276->10.217.0.201:8775: read: connection reset by peer" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.184395 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": read tcp 10.217.0.2:37260->10.217.0.201:8775: read: connection reset by peer" Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.392005 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.398605 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.400294 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.400329 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.508023 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerStarted","Data":"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.529008 4959 generic.go:334] "Generic (PLEG): container finished" podID="1da07093-5ea1-464a-a806-d30494ae9929" containerID="41f0dbd820ad59e9479c3ed201b87dee36f9fad0fd1f2a3b8f965b066cdb290a" exitCode=1 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.529582 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6565-account-delete-q7tmg" event={"ID":"1da07093-5ea1-464a-a806-d30494ae9929","Type":"ContainerDied","Data":"41f0dbd820ad59e9479c3ed201b87dee36f9fad0fd1f2a3b8f965b066cdb290a"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.542716 4959 generic.go:334] "Generic (PLEG): container finished" podID="3e788618-6d73-4794-8b04-ef41f70a94ab" containerID="55c2efbebd0db927e2a324867a441f257d17129410fe8d5306251eed8adffcef" exitCode=1 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.542781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef0b-account-delete-qnmr7" event={"ID":"3e788618-6d73-4794-8b04-ef41f70a94ab","Type":"ContainerDied","Data":"55c2efbebd0db927e2a324867a441f257d17129410fe8d5306251eed8adffcef"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.627831 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d265899-c75a-462f-9779-c573bd35a8d4" containerID="cfb72705ebccca6e3af246841cf201b6036478e1d451c879eb40ab8cf65d5799" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.627938 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerDied","Data":"cfb72705ebccca6e3af246841cf201b6036478e1d451c879eb40ab8cf65d5799"} Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.636712 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a is running failed: container process not found" containerID="1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.651459 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a is running failed: container process not found" containerID="1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.686459 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a is running failed: container process not found" containerID="1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.686510 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerName="nova-scheduler-scheduler" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.713558 4959 generic.go:334] "Generic (PLEG): container finished" podID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerID="ada15b2c15fe2f2dacc57422fd81cb3d234b56614971b21c30ab71da932c7b32" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.713588 4959 generic.go:334] "Generic (PLEG): container finished" podID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerID="13ffbebcbaac00743c439cf6bbe0144cf9261560d362f82d27b626d91213b204" exitCode=2 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.713595 4959 generic.go:334] "Generic (PLEG): container finished" podID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerID="4ddb932943c117d08c1006b64f264db30d2dae0973fb4d5017d89ad84b3ec3a8" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.715393 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363fcee0-62aa-4823-a5c8-c96c114bdeee" path="/var/lib/kubelet/pods/363fcee0-62aa-4823-a5c8-c96c114bdeee/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.715855 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43dd7882-778f-40c0-b7d4-757f090605d3" path="/var/lib/kubelet/pods/43dd7882-778f-40c0-b7d4-757f090605d3/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.721625 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" path="/var/lib/kubelet/pods/5dbba504-e66c-4dbe-abe3-4e69ceec1a23/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.722332 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2" path="/var/lib/kubelet/pods/6bf9bc8c-b8cf-49a7-bdcc-3758f7c066a2/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.722812 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d640be9-09ac-44ce-af77-68ea96aa6483" path="/var/lib/kubelet/pods/6d640be9-09ac-44ce-af77-68ea96aa6483/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.723892 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" path="/var/lib/kubelet/pods/ae7bb6a4-7314-482c-a09a-cb25b68d7e0a/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.724397 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce4b7381-3d0c-4067-a897-9b383b27a77c" path="/var/lib/kubelet/pods/ce4b7381-3d0c-4067-a897-9b383b27a77c/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.724893 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4495d5-4d1f-4ec5-af52-13a6a444148a" path="/var/lib/kubelet/pods/db4495d5-4d1f-4ec5-af52-13a6a444148a/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.726321 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" path="/var/lib/kubelet/pods/f325dd29-3b0d-4188-9ac2-e681d56e881f/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.726854 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb01d41b-854a-4072-8b05-807edc22812f" path="/var/lib/kubelet/pods/fb01d41b-854a-4072-8b05-807edc22812f/volumes" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.727674 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerDied","Data":"ada15b2c15fe2f2dacc57422fd81cb3d234b56614971b21c30ab71da932c7b32"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.727700 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerDied","Data":"13ffbebcbaac00743c439cf6bbe0144cf9261560d362f82d27b626d91213b204"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.727710 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerDied","Data":"4ddb932943c117d08c1006b64f264db30d2dae0973fb4d5017d89ad84b3ec3a8"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.748420 4959 generic.go:334] "Generic (PLEG): container finished" podID="d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" containerID="a8c979b1746ba8dacc24c61273c411303183dd2bb1f9d4920d65bb406dad43c5" exitCode=1 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.748503 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3e00-account-delete-qthlw" event={"ID":"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b","Type":"ContainerDied","Data":"a8c979b1746ba8dacc24c61273c411303183dd2bb1f9d4920d65bb406dad43c5"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.774299 4959 generic.go:334] "Generic (PLEG): container finished" podID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerID="fc9d972ac9b2f06f0e0356aa259d27ac5d465cfc53bdfcc611986b0718292921" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.774389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerDied","Data":"fc9d972ac9b2f06f0e0356aa259d27ac5d465cfc53bdfcc611986b0718292921"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.774444 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adc95387-65d6-40c6-82fb-e9a7f20bc514","Type":"ContainerDied","Data":"924b322a592049951ad85ab4516707758a112c97bd941a0fd5404e90ff4ca8c5"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.774455 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="924b322a592049951ad85ab4516707758a112c97bd941a0fd5404e90ff4ca8c5" Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.801187 4959 generic.go:334] "Generic (PLEG): container finished" podID="2ec784f6-4b46-435e-af12-e899033f8348" containerID="48413a96dbd55c4eac4e40c0f894a4fe721bcad741ce542a08321841cf77afd6" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.801491 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerDied","Data":"48413a96dbd55c4eac4e40c0f894a4fe721bcad741ce542a08321841cf77afd6"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.862442 4959 generic.go:334] "Generic (PLEG): container finished" podID="53ba77ff-6205-4215-bb28-a45ea0b7b39f" containerID="6a87988298ce622eae57f154f644d7de59597c62074a0c30695ef873834d9fc0" exitCode=1 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.862540 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2a29-account-delete-fhcv4" event={"ID":"53ba77ff-6205-4215-bb28-a45ea0b7b39f","Type":"ContainerDied","Data":"6a87988298ce622eae57f154f644d7de59597c62074a0c30695ef873834d9fc0"} Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.923374 4959 generic.go:334] "Generic (PLEG): container finished" podID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerID="1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.923461 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"632a7ace-4d5d-4588-af93-e0ab25125a14","Type":"ContainerDied","Data":"1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a"} Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.948463 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:31 crc kubenswrapper[4959]: E1003 13:56:31.985877 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.992172 4959 generic.go:334] "Generic (PLEG): container finished" podID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerID="b304c4a31b60f0e9c502c5bfcdf56c0274ea52bcd6515a31c6e5973c7f3cb733" exitCode=0 Oct 03 13:56:31 crc kubenswrapper[4959]: I1003 13:56:31.992279 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerDied","Data":"b304c4a31b60f0e9c502c5bfcdf56c0274ea52bcd6515a31c6e5973c7f3cb733"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.006417 4959 generic.go:334] "Generic (PLEG): container finished" podID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerID="9bc9b789b1e410f14438dd275ab4df1b4886b1a5b73a1ba9e3ec567e82067001" exitCode=0 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.006521 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerDied","Data":"9bc9b789b1e410f14438dd275ab4df1b4886b1a5b73a1ba9e3ec567e82067001"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.006576 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff","Type":"ContainerDied","Data":"10e3f5d56da42bee734ca1370b0c5cbc404d0a15225c6755fdb4a9e65b84865a"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.006590 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10e3f5d56da42bee734ca1370b0c5cbc404d0a15225c6755fdb4a9e65b84865a" Oct 03 13:56:32 crc kubenswrapper[4959]: E1003 13:56:32.010996 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:32 crc kubenswrapper[4959]: E1003 13:56:32.011043 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerName="nova-cell1-conductor-conductor" Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.020315 4959 generic.go:334] "Generic (PLEG): container finished" podID="943c9966-87bc-44fb-a19f-4ae15db1bdc4" containerID="13f35655ec5a85e5aa845c09f3f121c12edd40f9d1186f6c00e0a06cb62b25d7" exitCode=1 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.020498 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementaa19-account-delete-twnzm" event={"ID":"943c9966-87bc-44fb-a19f-4ae15db1bdc4","Type":"ContainerDied","Data":"13f35655ec5a85e5aa845c09f3f121c12edd40f9d1186f6c00e0a06cb62b25d7"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.035024 4959 generic.go:334] "Generic (PLEG): container finished" podID="f870f35a-1174-4568-9abb-b63d381483a2" containerID="65ae3048e36c4052ae2f42bf6462f21a9255680d29afaf9afdfbca6644e45e3c" exitCode=1 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.035179 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercca9-account-delete-778b2" event={"ID":"f870f35a-1174-4568-9abb-b63d381483a2","Type":"ContainerDied","Data":"65ae3048e36c4052ae2f42bf6462f21a9255680d29afaf9afdfbca6644e45e3c"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.088329 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerStarted","Data":"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.088462 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-b7cd665dc-9szsx" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker-log" containerID="cri-o://07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7" gracePeriod=30 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.088652 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-b7cd665dc-9szsx" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker" containerID="cri-o://65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053" gracePeriod=30 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.146749 4959 generic.go:334] "Generic (PLEG): container finished" podID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerID="72a3e7929b0adc554a8bfd78df9f7708f0538f33a7bbf4518d4a306cb7e072a8" exitCode=0 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.146833 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerDied","Data":"72a3e7929b0adc554a8bfd78df9f7708f0538f33a7bbf4518d4a306cb7e072a8"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.159271 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-b7cd665dc-9szsx" podStartSLOduration=7.159250489 podStartE2EDuration="7.159250489s" podCreationTimestamp="2025-10-03 13:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:56:32.144184163 +0000 UTC m=+1561.347527580" watchObservedRunningTime="2025-10-03 13:56:32.159250489 +0000 UTC m=+1561.362593906" Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.162318 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerID="06d4a6beaf8c7baed60c86f5ef66a97b9dcd29bdf45ef077eba2ec59e7cf24e6" exitCode=0 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.162379 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerDied","Data":"06d4a6beaf8c7baed60c86f5ef66a97b9dcd29bdf45ef077eba2ec59e7cf24e6"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.183203 4959 generic.go:334] "Generic (PLEG): container finished" podID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerID="95dbbc2b264de16726fd5cf962dcf5b2ea23cce70901b014633e800857d2e6ac" exitCode=0 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.183413 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerDied","Data":"95dbbc2b264de16726fd5cf962dcf5b2ea23cce70901b014633e800857d2e6ac"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.215214 4959 generic.go:334] "Generic (PLEG): container finished" podID="593264f5-6188-4861-9d83-6918034eb46e" containerID="437ecf2e1e223998fb6b040d3dc342771553ad152a649a307ba005f342deed15" exitCode=2 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.215315 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"593264f5-6188-4861-9d83-6918034eb46e","Type":"ContainerDied","Data":"437ecf2e1e223998fb6b040d3dc342771553ad152a649a307ba005f342deed15"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.215347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"593264f5-6188-4861-9d83-6918034eb46e","Type":"ContainerDied","Data":"f4bde2bb7cc603e46e0fb2ffd5621aa392edc67c8c3939960711728173c6543b"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.215359 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4bde2bb7cc603e46e0fb2ffd5621aa392edc67c8c3939960711728173c6543b" Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.236352 4959 generic.go:334] "Generic (PLEG): container finished" podID="a576d69d-a298-42d5-9c41-de82172c7a32" containerID="e9a759ce3223705acf4e1e483d5369ad492fba545b667e79de858b249d48d6d2" exitCode=1 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.240322 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3cd4-account-delete-f4dn6" event={"ID":"a576d69d-a298-42d5-9c41-de82172c7a32","Type":"ContainerDied","Data":"e9a759ce3223705acf4e1e483d5369ad492fba545b667e79de858b249d48d6d2"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.261463 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-86678496f6-dwd5r" event={"ID":"3edab447-50e9-4fae-a95f-d5a7cb489424","Type":"ContainerDied","Data":"0640b5671c1ba9b42e33af591d1536c6dae27bd4d2642a6ca2f4607601a4d974"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.261501 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0640b5671c1ba9b42e33af591d1536c6dae27bd4d2642a6ca2f4607601a4d974" Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.287715 4959 generic.go:334] "Generic (PLEG): container finished" podID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerID="a7c1d4fb80394150b035616306b0249611184085fee535c8c9ed09e63d2e3572" exitCode=0 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.287805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerDied","Data":"a7c1d4fb80394150b035616306b0249611184085fee535c8c9ed09e63d2e3572"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.300437 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerStarted","Data":"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0"} Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.300618 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener-log" containerID="cri-o://dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb" gracePeriod=30 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.300695 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener" containerID="cri-o://3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0" gracePeriod=30 Oct 03 13:56:32 crc kubenswrapper[4959]: I1003 13:56:32.339466 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" podStartSLOduration=7.339445295 podStartE2EDuration="7.339445295s" podCreationTimestamp="2025-10-03 13:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:56:32.335297944 +0000 UTC m=+1561.538641361" watchObservedRunningTime="2025-10-03 13:56:32.339445295 +0000 UTC m=+1561.542788712" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.246985 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.247400 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data podName:d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2 nodeName:}" failed. No retries permitted until 2025-10-03 13:56:41.247380943 +0000 UTC m=+1570.450724360 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data") pod "rabbitmq-server-0" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2") : configmap "rabbitmq-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.314110 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d375128e-6fb8-4c35-9def-4bd67c206d39","Type":"ContainerDied","Data":"e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.314522 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e819fda644aab834a824d98598e4a1bce568d072c581245dbf63cfba4a36cd28" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.317619 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron3e00-account-delete-qthlw" event={"ID":"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b","Type":"ContainerDied","Data":"11a81c2ca063c7d4aff8760ea97e39f85321948f51643fd76af45f40c8334c6f"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.317683 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11a81c2ca063c7d4aff8760ea97e39f85321948f51643fd76af45f40c8334c6f" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.320499 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance6565-account-delete-q7tmg" event={"ID":"1da07093-5ea1-464a-a806-d30494ae9929","Type":"ContainerDied","Data":"5d8bbcf7de2d7125a34188d0b847f27d498b39bdb6213d69b5bd4d7bd26850c4"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.320530 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d8bbcf7de2d7125a34188d0b847f27d498b39bdb6213d69b5bd4d7bd26850c4" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.323412 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3cd4-account-delete-f4dn6" event={"ID":"a576d69d-a298-42d5-9c41-de82172c7a32","Type":"ContainerDied","Data":"a29734fee59c9beeb03f3ad6d1144d2d4658d86e65b681058373ba43b2de9a9b"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.323457 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a29734fee59c9beeb03f3ad6d1144d2d4658d86e65b681058373ba43b2de9a9b" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.325137 4959 generic.go:334] "Generic (PLEG): container finished" podID="30542159-188f-4508-a1b4-70285f59aef3" containerID="aae8036f06d6bb33643dcc5d309e687768c23fb0437205d1473dc1f70e1e4751" exitCode=0 Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.325228 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"30542159-188f-4508-a1b4-70285f59aef3","Type":"ContainerDied","Data":"aae8036f06d6bb33643dcc5d309e687768c23fb0437205d1473dc1f70e1e4751"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.325265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"30542159-188f-4508-a1b4-70285f59aef3","Type":"ContainerDied","Data":"5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.325279 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.326762 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican2a29-account-delete-fhcv4" event={"ID":"53ba77ff-6205-4215-bb28-a45ea0b7b39f","Type":"ContainerDied","Data":"33e104eb4372325d59ebd0f1bc19b917f3a71beed66561e03bdbe61fe52d1595"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.326784 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33e104eb4372325d59ebd0f1bc19b917f3a71beed66561e03bdbe61fe52d1595" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.331271 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-677d776894-ljrrj" event={"ID":"8d2a34ba-1b42-4651-93c7-8855f2398ff5","Type":"ContainerDied","Data":"b1768aacb21314d2a0749c15d34ef9d67cc08b934e893da49b8d325044b8c354"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.331296 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1768aacb21314d2a0749c15d34ef9d67cc08b934e893da49b8d325044b8c354" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.332981 4959 generic.go:334] "Generic (PLEG): container finished" podID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerID="e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" exitCode=0 Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.333063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ff0c0088-66f1-4f85-8a34-a1f34867a37a","Type":"ContainerDied","Data":"e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.333268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ff0c0088-66f1-4f85-8a34-a1f34867a37a","Type":"ContainerDied","Data":"043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.333361 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.335059 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" event={"ID":"2ec784f6-4b46-435e-af12-e899033f8348","Type":"ContainerDied","Data":"16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.335090 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16e38d4e54b08f1c18ee942c0df417d9a852275da4571c2718bbef059060b60c" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.336844 4959 generic.go:334] "Generic (PLEG): container finished" podID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerID="dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb" exitCode=143 Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.336904 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerDied","Data":"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.338439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercca9-account-delete-778b2" event={"ID":"f870f35a-1174-4568-9abb-b63d381483a2","Type":"ContainerDied","Data":"5c0c1f1ab4ccf63b22028b84710b27349347718809db875c9643fa860262fd5a"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.338531 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c0c1f1ab4ccf63b22028b84710b27349347718809db875c9643fa860262fd5a" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.343885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef0b-account-delete-qnmr7" event={"ID":"3e788618-6d73-4794-8b04-ef41f70a94ab","Type":"ContainerDied","Data":"8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.343927 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6d4154b0697ed6978937ff36f750d3be7c94fdf95758452342f9c0b17c97be" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.345572 4959 generic.go:334] "Generic (PLEG): container finished" podID="b963e179-793e-4441-9c4d-607ab547e0ea" containerID="07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7" exitCode=143 Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.345597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerDied","Data":"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.351249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"632a7ace-4d5d-4588-af93-e0ab25125a14","Type":"ContainerDied","Data":"e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.351285 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.353269 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-54b8ffdcf-7hs45" event={"ID":"aaa51dc6-b33c-41fc-9829-e4d20da77df3","Type":"ContainerDied","Data":"90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.353297 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.355089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementaa19-account-delete-twnzm" event={"ID":"943c9966-87bc-44fb-a19f-4ae15db1bdc4","Type":"ContainerDied","Data":"255a0bea8e5d0b5422082e03fdb0393292916b72890c8510403dc958dc9ecd28"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.355114 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="255a0bea8e5d0b5422082e03fdb0393292916b72890c8510403dc958dc9ecd28" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.360117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a","Type":"ContainerDied","Data":"1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.360144 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1583b0e5e64c689d057dfa001aa9f96f597610a00d004d49a5007b552c7b95f3" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.362389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d265899-c75a-462f-9779-c573bd35a8d4","Type":"ContainerDied","Data":"caae51897f61d45af2f7174fb8c8b4b34f4a757768c222ebc005f304dfb827c6"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.362429 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caae51897f61d45af2f7174fb8c8b4b34f4a757768c222ebc005f304dfb827c6" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.364895 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2f9c520-72f3-482f-be82-8a625a24f3e1","Type":"ContainerDied","Data":"e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2"} Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.364920 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7d03e3f4ed7a6ae9767a2c99b1fb808c6ecb7cb53f7c639cdb74209bf4bf7c2" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.378937 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.379265 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.379622 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.379664 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.381587 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.384059 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.385255 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.385295 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.492445 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.514952 4959 scope.go:117] "RemoveContainer" containerID="e51713b9b874f135af735bd09481738fbf57ba1d4691649ef78b87b0fca43cd1" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.537576 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551458 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551652 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551780 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbt7t\" (UniqueName: \"kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551821 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551879 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.551936 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle\") pod \"3edab447-50e9-4fae-a95f-d5a7cb489424\" (UID: \"3edab447-50e9-4fae-a95f-d5a7cb489424\") " Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.552557 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.552609 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:41.552592668 +0000 UTC m=+1570.755936085 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.556451 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs" (OuterVolumeSpecName: "logs") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.561822 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t" (OuterVolumeSpecName: "kube-api-access-mbt7t") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "kube-api-access-mbt7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.570848 4959 scope.go:117] "RemoveContainer" containerID="802921468f08beb7a7faecf994bffea767c35ceda75b15525c02a4e2c871e13a" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.574796 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts" (OuterVolumeSpecName: "scripts") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.576278 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.576690 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.578480 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.594267 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.610385 4959 scope.go:117] "RemoveContainer" containerID="e9f6f6f4ded33babe3c6b4e9a7384c3513fe53da441505d10b5d13c79bd8152e" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.611812 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.613131 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.616862 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.627342 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.627398 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656523 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656592 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656639 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqtlg\" (UniqueName: \"kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg\") pod \"593264f5-6188-4861-9d83-6918034eb46e\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656657 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6x7d\" (UniqueName: \"kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656678 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom\") pod \"2ec784f6-4b46-435e-af12-e899033f8348\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656721 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656768 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle\") pod \"593264f5-6188-4861-9d83-6918034eb46e\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656787 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs\") pod \"2ec784f6-4b46-435e-af12-e899033f8348\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data\") pod \"2ec784f6-4b46-435e-af12-e899033f8348\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656817 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656832 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656850 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656892 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656912 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle\") pod \"2ec784f6-4b46-435e-af12-e899033f8348\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656965 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656983 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.656998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d5fl\" (UniqueName: \"kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657016 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657038 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs\") pod \"593264f5-6188-4861-9d83-6918034eb46e\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657055 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config\") pod \"593264f5-6188-4861-9d83-6918034eb46e\" (UID: \"593264f5-6188-4861-9d83-6918034eb46e\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657077 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657092 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs\") pod \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\" (UID: \"71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657110 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657160 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657181 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657215 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657230 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657253 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw5gn\" (UniqueName: \"kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn\") pod \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\" (UID: \"33410f10-ea8e-4ee3-be69-01f1cb9fe9ff\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657275 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8wr\" (UniqueName: \"kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr\") pod \"2ec784f6-4b46-435e-af12-e899033f8348\" (UID: \"2ec784f6-4b46-435e-af12-e899033f8348\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657322 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs\") pod \"adc95387-65d6-40c6-82fb-e9a7f20bc514\" (UID: \"adc95387-65d6-40c6-82fb-e9a7f20bc514\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657852 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3edab447-50e9-4fae-a95f-d5a7cb489424-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657871 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbt7t\" (UniqueName: \"kubernetes.io/projected/3edab447-50e9-4fae-a95f-d5a7cb489424-kube-api-access-mbt7t\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.657881 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.661639 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs" (OuterVolumeSpecName: "logs") pod "2ec784f6-4b46-435e-af12-e899033f8348" (UID: "2ec784f6-4b46-435e-af12-e899033f8348"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.662792 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs" (OuterVolumeSpecName: "logs") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.669308 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.671273 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.672501 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.674785 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs" (OuterVolumeSpecName: "logs") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.680247 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs" (OuterVolumeSpecName: "logs") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.696040 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg" (OuterVolumeSpecName: "kube-api-access-fqtlg") pod "593264f5-6188-4861-9d83-6918034eb46e" (UID: "593264f5-6188-4861-9d83-6918034eb46e"). InnerVolumeSpecName "kube-api-access-fqtlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.710567 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.717224 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.718670 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.718703 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.720177 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.720261 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.720315 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2ec784f6-4b46-435e-af12-e899033f8348" (UID: "2ec784f6-4b46-435e-af12-e899033f8348"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.720336 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn" (OuterVolumeSpecName: "kube-api-access-hw5gn") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "kube-api-access-hw5gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.720655 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts" (OuterVolumeSpecName: "scripts") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.723493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl" (OuterVolumeSpecName: "kube-api-access-2d5fl") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "kube-api-access-2d5fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.740648 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d" (OuterVolumeSpecName: "kube-api-access-m6x7d") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "kube-api-access-m6x7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.746938 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.747016 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr" (OuterVolumeSpecName: "kube-api-access-qp8wr") pod "2ec784f6-4b46-435e-af12-e899033f8348" (UID: "2ec784f6-4b46-435e-af12-e899033f8348"). InnerVolumeSpecName "kube-api-access-qp8wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.747052 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts" (OuterVolumeSpecName: "scripts") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.748328 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts" (OuterVolumeSpecName: "scripts") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.771615 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.771676 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.771704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.771885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f6gf\" (UniqueName: \"kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.771976 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772007 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id\") pod \"f2f9c520-72f3-482f-be82-8a625a24f3e1\" (UID: \"f2f9c520-72f3-482f-be82-8a625a24f3e1\") " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772589 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772607 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqtlg\" (UniqueName: \"kubernetes.io/projected/593264f5-6188-4861-9d83-6918034eb46e-kube-api-access-fqtlg\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772620 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6x7d\" (UniqueName: \"kubernetes.io/projected/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-kube-api-access-m6x7d\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772633 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772643 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772654 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ec784f6-4b46-435e-af12-e899033f8348-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772676 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772688 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772699 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adc95387-65d6-40c6-82fb-e9a7f20bc514-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772709 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772721 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d5fl\" (UniqueName: \"kubernetes.io/projected/adc95387-65d6-40c6-82fb-e9a7f20bc514-kube-api-access-2d5fl\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772731 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772741 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772758 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772769 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772779 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772793 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw5gn\" (UniqueName: \"kubernetes.io/projected/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-kube-api-access-hw5gn\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772805 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8wr\" (UniqueName: \"kubernetes.io/projected/2ec784f6-4b46-435e-af12-e899033f8348-kube-api-access-qp8wr\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.772818 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.776398 4959 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: E1003 13:56:33.776445 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data podName:1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d nodeName:}" failed. No retries permitted until 2025-10-03 13:56:41.776431238 +0000 UTC m=+1570.979774655 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data") pod "rabbitmq-cell1-server-0" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d") : configmap "rabbitmq-cell1-config-data" not found Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.776759 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.788220 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.815618 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts" (OuterVolumeSpecName: "scripts") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.815669 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf" (OuterVolumeSpecName: "kube-api-access-8f6gf") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "kube-api-access-8f6gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.838695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.874412 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.874433 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.874444 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f6gf\" (UniqueName: \"kubernetes.io/projected/f2f9c520-72f3-482f-be82-8a625a24f3e1-kube-api-access-8f6gf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.874453 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2f9c520-72f3-482f-be82-8a625a24f3e1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.966065 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:56:33 crc kubenswrapper[4959]: I1003 13:56:33.978374 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.099883 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data" (OuterVolumeSpecName: "config-data") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.129713 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.148969 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.166731 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" probeResult="failure" output=< Oct 03 13:56:34 crc kubenswrapper[4959]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 03 13:56:34 crc kubenswrapper[4959]: > Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.187643 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.189587 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.189628 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.189639 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.189648 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.197187 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "593264f5-6188-4861-9d83-6918034eb46e" (UID: "593264f5-6188-4861-9d83-6918034eb46e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.200566 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "593264f5-6188-4861-9d83-6918034eb46e" (UID: "593264f5-6188-4861-9d83-6918034eb46e"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.215812 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ec784f6-4b46-435e-af12-e899033f8348" (UID: "2ec784f6-4b46-435e-af12-e899033f8348"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.221829 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.225361 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.265540 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.283326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "adc95387-65d6-40c6-82fb-e9a7f20bc514" (UID: "adc95387-65d6-40c6-82fb-e9a7f20bc514"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291776 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291811 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc95387-65d6-40c6-82fb-e9a7f20bc514-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291823 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291834 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291845 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291855 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.291866 4959 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.308058 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.315960 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "593264f5-6188-4861-9d83-6918034eb46e" (UID: "593264f5-6188-4861-9d83-6918034eb46e"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.339338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.341914 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data" (OuterVolumeSpecName: "config-data") pod "2ec784f6-4b46-435e-af12-e899033f8348" (UID: "2ec784f6-4b46-435e-af12-e899033f8348"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.343348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data" (OuterVolumeSpecName: "config-data") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.353356 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.368854 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data" (OuterVolumeSpecName: "config-data") pod "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" (UID: "33410f10-ea8e-4ee3-be69-01f1cb9fe9ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.374675 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data" (OuterVolumeSpecName: "config-data") pod "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" (UID: "71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.384367 4959 generic.go:334] "Generic (PLEG): container finished" podID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerID="73f4e64df58083370f19d3a84fd7e6d994163b876b512f98d2ee1e62e5754b19" exitCode=0 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.388853 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.390068 4959 generic.go:334] "Generic (PLEG): container finished" podID="f51595d7-80bf-4436-9429-9f0da918850d" containerID="964de272e99fd5d3325f928f234440a05a23315ec4d4c92db1d5a71e50facdcf" exitCode=0 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.393095 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" containerID="cri-o://bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a" gracePeriod=30 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.393185 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" containerID="cri-o://659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879" gracePeriod=30 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394867 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394893 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394901 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394911 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ec784f6-4b46-435e-af12-e899033f8348-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394922 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394929 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394938 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394945 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.394954 4959 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/593264f5-6188-4861-9d83-6918034eb46e-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.396671 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3edab447-50e9-4fae-a95f-d5a7cb489424" (UID: "3edab447-50e9-4fae-a95f-d5a7cb489424"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.397105 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ade995a8-1a92-43c8-9018-583d339f5f01/ovn-northd/0.log" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.397141 4959 generic.go:334] "Generic (PLEG): container finished" podID="ade995a8-1a92-43c8-9018-583d339f5f01" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" exitCode=139 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.401695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data" (OuterVolumeSpecName: "config-data") pod "f2f9c520-72f3-482f-be82-8a625a24f3e1" (UID: "f2f9c520-72f3-482f-be82-8a625a24f3e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403551 4959 generic.go:334] "Generic (PLEG): container finished" podID="9ceadbac-10d3-476b-8250-5005c4391e71" containerID="f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" exitCode=0 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403640 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f558c656-kjrrw" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403653 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403777 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403816 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403849 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.403879 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-86678496f6-dwd5r" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.404004 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494104 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerDied","Data":"73f4e64df58083370f19d3a84fd7e6d994163b876b512f98d2ee1e62e5754b19"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2","Type":"ContainerDied","Data":"24c85da72d38b0d376a5a5ae6ab49af74c8e798db542889a1745ed029c262fb2"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494164 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24c85da72d38b0d376a5a5ae6ab49af74c8e798db542889a1745ed029c262fb2" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494180 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494194 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494231 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b6c4d45db-cwzjn" event={"ID":"f51595d7-80bf-4436-9429-9f0da918850d","Type":"ContainerDied","Data":"964de272e99fd5d3325f928f234440a05a23315ec4d4c92db1d5a71e50facdcf"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494248 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerStarted","Data":"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerDied","Data":"7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494269 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ade995a8-1a92-43c8-9018-583d339f5f01","Type":"ContainerDied","Data":"d25d5d7ca31fbf65ba57aee081929fb858eb137519251181c4cf138741675524"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494277 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d25d5d7ca31fbf65ba57aee081929fb858eb137519251181c4cf138741675524" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494285 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerDied","Data":"f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494296 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ceadbac-10d3-476b-8250-5005c4391e71","Type":"ContainerDied","Data":"46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c"} Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.494310 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.497599 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2f9c520-72f3-482f-be82-8a625a24f3e1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.497632 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3edab447-50e9-4fae-a95f-d5a7cb489424-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.515259 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.537131 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-544d968d58-6cpgj" podStartSLOduration=9.537110554 podStartE2EDuration="9.537110554s" podCreationTimestamp="2025-10-03 13:56:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:56:34.417965287 +0000 UTC m=+1563.621308704" watchObservedRunningTime="2025-10-03 13:56:34.537110554 +0000 UTC m=+1563.740453971" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598517 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598583 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598696 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598727 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598747 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.598799 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d5wz\" (UniqueName: \"kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz\") pod \"d375128e-6fb8-4c35-9def-4bd67c206d39\" (UID: \"d375128e-6fb8-4c35-9def-4bd67c206d39\") " Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.599823 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs" (OuterVolumeSpecName: "logs") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.606375 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz" (OuterVolumeSpecName: "kube-api-access-9d5wz") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "kube-api-access-9d5wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.642656 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data" (OuterVolumeSpecName: "config-data") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.644232 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.674388 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.689573 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:56:34 crc kubenswrapper[4959]: E1003 13:56:34.689896 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.701175 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.701218 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d375128e-6fb8-4c35-9def-4bd67c206d39-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.701227 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.701238 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.701247 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d5wz\" (UniqueName: \"kubernetes.io/projected/d375128e-6fb8-4c35-9def-4bd67c206d39-kube-api-access-9d5wz\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.729333 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d375128e-6fb8-4c35-9def-4bd67c206d39" (UID: "d375128e-6fb8-4c35-9def-4bd67c206d39"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:34 crc kubenswrapper[4959]: E1003 13:56:34.764506 4959 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 03 13:56:34 crc kubenswrapper[4959]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-03T13:56:27Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 03 13:56:34 crc kubenswrapper[4959]: /etc/init.d/functions: line 589: 463 Alarm clock "$@" Oct 03 13:56:34 crc kubenswrapper[4959]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-k4vvh" message=< Oct 03 13:56:34 crc kubenswrapper[4959]: Exiting ovn-controller (1) [FAILED] Oct 03 13:56:34 crc kubenswrapper[4959]: Killing ovn-controller (1) [ OK ] Oct 03 13:56:34 crc kubenswrapper[4959]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 03 13:56:34 crc kubenswrapper[4959]: 2025-10-03T13:56:27Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 03 13:56:34 crc kubenswrapper[4959]: /etc/init.d/functions: line 589: 463 Alarm clock "$@" Oct 03 13:56:34 crc kubenswrapper[4959]: > Oct 03 13:56:34 crc kubenswrapper[4959]: E1003 13:56:34.764543 4959 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 03 13:56:34 crc kubenswrapper[4959]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-03T13:56:27Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 03 13:56:34 crc kubenswrapper[4959]: /etc/init.d/functions: line 589: 463 Alarm clock "$@" Oct 03 13:56:34 crc kubenswrapper[4959]: > pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" containerID="cri-o://d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.764585 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-k4vvh" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" containerID="cri-o://d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119" gracePeriod=22 Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.802568 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d375128e-6fb8-4c35-9def-4bd67c206d39-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:34 crc kubenswrapper[4959]: I1003 13:56:34.912364 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-76dd7cdf5-g59pr" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": dial tcp 10.217.0.158:9696: connect: connection refused" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.045651 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.069280 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.082462 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.088370 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.092453 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.092858 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.098960 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.104914 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.105959 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.110876 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcmjk\" (UniqueName: \"kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk\") pod \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.111290 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data\") pod \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.111507 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs\") pod \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.111623 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom\") pod \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.111770 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle\") pod \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\" (UID: \"aaa51dc6-b33c-41fc-9829-e4d20da77df3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.112115 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs" (OuterVolumeSpecName: "logs") pod "aaa51dc6-b33c-41fc-9829-e4d20da77df3" (UID: "aaa51dc6-b33c-41fc-9829-e4d20da77df3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.119817 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.126853 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.136153 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.139099 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.155906 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.163084 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.183552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.190637 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.195800 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7f558c656-kjrrw"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.206802 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.207719 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.209437 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.214267 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs\") pod \"4d265899-c75a-462f-9779-c573bd35a8d4\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.214882 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215404 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215494 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle\") pod \"4d265899-c75a-462f-9779-c573bd35a8d4\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215584 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215651 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgwq7\" (UniqueName: \"kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7\") pod \"53ba77ff-6205-4215-bb28-a45ea0b7b39f\" (UID: \"53ba77ff-6205-4215-bb28-a45ea0b7b39f\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215819 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ck75\" (UniqueName: \"kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215910 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh78d\" (UniqueName: \"kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d\") pod \"4d265899-c75a-462f-9779-c573bd35a8d4\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215927 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs\") pod \"4d265899-c75a-462f-9779-c573bd35a8d4\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.215995 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216018 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x8qd\" (UniqueName: \"kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd\") pod \"3e788618-6d73-4794-8b04-ef41f70a94ab\" (UID: \"3e788618-6d73-4794-8b04-ef41f70a94ab\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216097 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jnfg\" (UniqueName: \"kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg\") pod \"1da07093-5ea1-464a-a806-d30494ae9929\" (UID: \"1da07093-5ea1-464a-a806-d30494ae9929\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216158 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86czq\" (UniqueName: \"kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq\") pod \"943c9966-87bc-44fb-a19f-4ae15db1bdc4\" (UID: \"943c9966-87bc-44fb-a19f-4ae15db1bdc4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216597 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216636 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfwd7\" (UniqueName: \"kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7\") pod \"632a7ace-4d5d-4588-af93-e0ab25125a14\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216676 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data\") pod \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\" (UID: \"8d2a34ba-1b42-4651-93c7-8855f2398ff5\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216749 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xckqp\" (UniqueName: \"kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp\") pod \"f870f35a-1174-4568-9abb-b63d381483a2\" (UID: \"f870f35a-1174-4568-9abb-b63d381483a2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216773 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data\") pod \"632a7ace-4d5d-4588-af93-e0ab25125a14\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216834 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle\") pod \"632a7ace-4d5d-4588-af93-e0ab25125a14\" (UID: \"632a7ace-4d5d-4588-af93-e0ab25125a14\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.216870 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data\") pod \"4d265899-c75a-462f-9779-c573bd35a8d4\" (UID: \"4d265899-c75a-462f-9779-c573bd35a8d4\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.217723 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs" (OuterVolumeSpecName: "logs") pod "4d265899-c75a-462f-9779-c573bd35a8d4" (UID: "4d265899-c75a-462f-9779-c573bd35a8d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.220668 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg" (OuterVolumeSpecName: "kube-api-access-7jnfg") pod "1da07093-5ea1-464a-a806-d30494ae9929" (UID: "1da07093-5ea1-464a-a806-d30494ae9929"). InnerVolumeSpecName "kube-api-access-7jnfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.221209 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.225988 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d" (OuterVolumeSpecName: "kube-api-access-lh78d") pod "4d265899-c75a-462f-9779-c573bd35a8d4" (UID: "4d265899-c75a-462f-9779-c573bd35a8d4"). InnerVolumeSpecName "kube-api-access-lh78d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.226954 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.228444 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp" (OuterVolumeSpecName: "kube-api-access-xckqp") pod "f870f35a-1174-4568-9abb-b63d381483a2" (UID: "f870f35a-1174-4568-9abb-b63d381483a2"). InnerVolumeSpecName "kube-api-access-xckqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.233938 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaa51dc6-b33c-41fc-9829-e4d20da77df3" (UID: "aaa51dc6-b33c-41fc-9829-e4d20da77df3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.234222 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd" (OuterVolumeSpecName: "kube-api-access-4x8qd") pod "3e788618-6d73-4794-8b04-ef41f70a94ab" (UID: "3e788618-6d73-4794-8b04-ef41f70a94ab"). InnerVolumeSpecName "kube-api-access-4x8qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.234308 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75" (OuterVolumeSpecName: "kube-api-access-9ck75") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "kube-api-access-9ck75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.234426 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data" (OuterVolumeSpecName: "config-data") pod "aaa51dc6-b33c-41fc-9829-e4d20da77df3" (UID: "aaa51dc6-b33c-41fc-9829-e4d20da77df3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.234693 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.237946 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d265899-c75a-462f-9779-c573bd35a8d4-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.237972 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ck75\" (UniqueName: \"kubernetes.io/projected/8d2a34ba-1b42-4651-93c7-8855f2398ff5-kube-api-access-9ck75\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.237983 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh78d\" (UniqueName: \"kubernetes.io/projected/4d265899-c75a-462f-9779-c573bd35a8d4-kube-api-access-lh78d\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.237993 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x8qd\" (UniqueName: \"kubernetes.io/projected/3e788618-6d73-4794-8b04-ef41f70a94ab-kube-api-access-4x8qd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238002 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jnfg\" (UniqueName: \"kubernetes.io/projected/1da07093-5ea1-464a-a806-d30494ae9929-kube-api-access-7jnfg\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238010 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238018 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa51dc6-b33c-41fc-9829-e4d20da77df3-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238029 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xckqp\" (UniqueName: \"kubernetes.io/projected/f870f35a-1174-4568-9abb-b63d381483a2-kube-api-access-xckqp\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238037 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.238415 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs" (OuterVolumeSpecName: "logs") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.242298 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.248130 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.250603 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7" (OuterVolumeSpecName: "kube-api-access-lfwd7") pod "632a7ace-4d5d-4588-af93-e0ab25125a14" (UID: "632a7ace-4d5d-4588-af93-e0ab25125a14"). InnerVolumeSpecName "kube-api-access-lfwd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.250911 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.251271 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7" (OuterVolumeSpecName: "kube-api-access-rgwq7") pod "53ba77ff-6205-4215-bb28-a45ea0b7b39f" (UID: "53ba77ff-6205-4215-bb28-a45ea0b7b39f"). InnerVolumeSpecName "kube-api-access-rgwq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.251868 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.255517 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aaa51dc6-b33c-41fc-9829-e4d20da77df3" (UID: "aaa51dc6-b33c-41fc-9829-e4d20da77df3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.260706 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.267267 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-86678496f6-dwd5r"] Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.268383 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq" (OuterVolumeSpecName: "kube-api-access-86czq") pod "943c9966-87bc-44fb-a19f-4ae15db1bdc4" (UID: "943c9966-87bc-44fb-a19f-4ae15db1bdc4"). InnerVolumeSpecName "kube-api-access-86czq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.269607 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ade995a8-1a92-43c8-9018-583d339f5f01/ovn-northd/0.log" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.269865 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.285568 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.312439 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk" (OuterVolumeSpecName: "kube-api-access-wcmjk") pod "aaa51dc6-b33c-41fc-9829-e4d20da77df3" (UID: "aaa51dc6-b33c-41fc-9829-e4d20da77df3"). InnerVolumeSpecName "kube-api-access-wcmjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.339540 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340136 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340181 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle\") pod \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340364 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk544\" (UniqueName: \"kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544\") pod \"a576d69d-a298-42d5-9c41-de82172c7a32\" (UID: \"a576d69d-a298-42d5-9c41-de82172c7a32\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340396 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340417 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340460 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340475 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config\") pod \"30542159-188f-4508-a1b4-70285f59aef3\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340596 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data\") pod \"30542159-188f-4508-a1b4-70285f59aef3\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340629 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lkzg\" (UniqueName: \"kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg\") pod \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dngf\" (UniqueName: \"kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf\") pod \"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b\" (UID: \"d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340707 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.340722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmjbz\" (UniqueName: \"kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351363 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data\") pod \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\" (UID: \"ff0c0088-66f1-4f85-8a34-a1f34867a37a\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351432 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle\") pod \"30542159-188f-4508-a1b4-70285f59aef3\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351514 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs\") pod \"30542159-188f-4508-a1b4-70285f59aef3\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351554 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts\") pod \"9ceadbac-10d3-476b-8250-5005c4391e71\" (UID: \"9ceadbac-10d3-476b-8250-5005c4391e71\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.351581 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsh8j\" (UniqueName: \"kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j\") pod \"30542159-188f-4508-a1b4-70285f59aef3\" (UID: \"30542159-188f-4508-a1b4-70285f59aef3\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352232 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352249 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgwq7\" (UniqueName: \"kubernetes.io/projected/53ba77ff-6205-4215-bb28-a45ea0b7b39f-kube-api-access-rgwq7\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352260 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcmjk\" (UniqueName: \"kubernetes.io/projected/aaa51dc6-b33c-41fc-9829-e4d20da77df3-kube-api-access-wcmjk\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352273 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86czq\" (UniqueName: \"kubernetes.io/projected/943c9966-87bc-44fb-a19f-4ae15db1bdc4-kube-api-access-86czq\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352284 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d2a34ba-1b42-4651-93c7-8855f2398ff5-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352295 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfwd7\" (UniqueName: \"kubernetes.io/projected/632a7ace-4d5d-4588-af93-e0ab25125a14-kube-api-access-lfwd7\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.352307 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa51dc6-b33c-41fc-9829-e4d20da77df3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.363562 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d265899-c75a-462f-9779-c573bd35a8d4" (UID: "4d265899-c75a-462f-9779-c573bd35a8d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.374776 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.378381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.389975 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.391476 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz" (OuterVolumeSpecName: "kube-api-access-lmjbz") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "kube-api-access-lmjbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.399879 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg" (OuterVolumeSpecName: "kube-api-access-8lkzg") pod "ff0c0088-66f1-4f85-8a34-a1f34867a37a" (UID: "ff0c0088-66f1-4f85-8a34-a1f34867a37a"). InnerVolumeSpecName "kube-api-access-8lkzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.414506 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.428573 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544" (OuterVolumeSpecName: "kube-api-access-qk544") pod "a576d69d-a298-42d5-9c41-de82172c7a32" (UID: "a576d69d-a298-42d5-9c41-de82172c7a32"). InnerVolumeSpecName "kube-api-access-qk544". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.444362 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf" (OuterVolumeSpecName: "kube-api-access-8dngf") pod "d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" (UID: "d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b"). InnerVolumeSpecName "kube-api-access-8dngf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461233 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461294 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461326 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461373 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461447 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461495 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461530 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461554 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461593 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461621 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461642 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461671 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgbm9\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461700 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461722 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461756 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461812 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461844 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461867 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461909 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461949 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.461985 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462026 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462076 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462100 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462126 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd8ph\" (UniqueName: \"kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462213 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462246 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462286 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462323 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462355 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg86x\" (UniqueName: \"kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x\") pod \"ade995a8-1a92-43c8-9018-583d339f5f01\" (UID: \"ade995a8-1a92-43c8-9018-583d339f5f01\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462377 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462416 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462447 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf\") pod \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\" (UID: \"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462477 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys\") pod \"f51595d7-80bf-4436-9429-9f0da918850d\" (UID: \"f51595d7-80bf-4436-9429-9f0da918850d\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462499 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwxb8\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8\") pod \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\" (UID: \"d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462899 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk544\" (UniqueName: \"kubernetes.io/projected/a576d69d-a298-42d5-9c41-de82172c7a32-kube-api-access-qk544\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462915 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462928 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lkzg\" (UniqueName: \"kubernetes.io/projected/ff0c0088-66f1-4f85-8a34-a1f34867a37a-kube-api-access-8lkzg\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462941 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dngf\" (UniqueName: \"kubernetes.io/projected/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b-kube-api-access-8dngf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462953 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmjbz\" (UniqueName: \"kubernetes.io/projected/9ceadbac-10d3-476b-8250-5005c4391e71-kube-api-access-lmjbz\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462965 4959 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462976 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.462988 4959 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.463000 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ceadbac-10d3-476b-8250-5005c4391e71-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.464699 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets" (OuterVolumeSpecName: "secrets") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.469341 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.472562 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.472577 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.490446 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b6c4d45db-cwzjn" event={"ID":"f51595d7-80bf-4436-9429-9f0da918850d","Type":"ContainerDied","Data":"e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.490496 4959 scope.go:117] "RemoveContainer" containerID="964de272e99fd5d3325f928f234440a05a23315ec4d4c92db1d5a71e50facdcf" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.490606 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b6c4d45db-cwzjn" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.493556 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "30542159-188f-4508-a1b4-70285f59aef3" (UID: "30542159-188f-4508-a1b4-70285f59aef3"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.493762 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.494074 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.501450 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j" (OuterVolumeSpecName: "kube-api-access-nsh8j") pod "30542159-188f-4508-a1b4-70285f59aef3" (UID: "30542159-188f-4508-a1b4-70285f59aef3"). InnerVolumeSpecName "kube-api-access-nsh8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.501597 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data" (OuterVolumeSpecName: "config-data") pod "30542159-188f-4508-a1b4-70285f59aef3" (UID: "30542159-188f-4508-a1b4-70285f59aef3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.501839 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.502263 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.511374 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config" (OuterVolumeSpecName: "config") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.523791 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph" (OuterVolumeSpecName: "kube-api-access-jd8ph") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "kube-api-access-jd8ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.529972 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8" (OuterVolumeSpecName: "kube-api-access-mwxb8") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "kube-api-access-mwxb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.531373 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts" (OuterVolumeSpecName: "scripts") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537425 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts" (OuterVolumeSpecName: "scripts") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537525 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537779 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537871 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info" (OuterVolumeSpecName: "pod-info") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537885 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.537946 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info" (OuterVolumeSpecName: "pod-info") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.538878 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.547070 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.547152 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.547425 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9" (OuterVolumeSpecName: "kube-api-access-jgbm9") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "kube-api-access-jgbm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.547543 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.548858 4959 generic.go:334] "Generic (PLEG): container finished" podID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerID="47160c83891466f0f38183ffc99e4cb435c5e8b448cf8d1325e15334ebeb44d0" exitCode=0 Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.548919 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerDied","Data":"47160c83891466f0f38183ffc99e4cb435c5e8b448cf8d1325e15334ebeb44d0"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.554042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x" (OuterVolumeSpecName: "kube-api-access-sg86x") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "kube-api-access-sg86x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.559886 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568575 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg86x\" (UniqueName: \"kubernetes.io/projected/ade995a8-1a92-43c8-9018-583d339f5f01-kube-api-access-sg86x\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568611 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568624 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568636 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568650 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwxb8\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-kube-api-access-mwxb8\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568660 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568670 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568681 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsh8j\" (UniqueName: \"kubernetes.io/projected/30542159-188f-4508-a1b4-70285f59aef3-kube-api-access-nsh8j\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568691 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568702 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568712 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568724 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ade995a8-1a92-43c8-9018-583d339f5f01-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568734 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568746 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgbm9\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-kube-api-access-jgbm9\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568758 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568769 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568780 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568790 4959 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568824 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568836 4959 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568852 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568867 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568879 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568896 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568907 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568919 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30542159-188f-4508-a1b4-70285f59aef3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568930 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.568943 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd8ph\" (UniqueName: \"kubernetes.io/projected/f51595d7-80bf-4436-9429-9f0da918850d-kube-api-access-jd8ph\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.572287 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.594509 4959 generic.go:334] "Generic (PLEG): container finished" podID="71930c60-fa28-4810-98ca-6406778f917e" containerID="bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a" exitCode=143 Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.594867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerDied","Data":"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.600186 4959 generic.go:334] "Generic (PLEG): container finished" podID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerID="79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6" exitCode=0 Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.600319 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.600314 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerDied","Data":"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.600369 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d","Type":"ContainerDied","Data":"8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.600393 4959 scope.go:117] "RemoveContainer" containerID="79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.601077 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k4vvh_8729e6fd-0547-43a9-bbef-d4dcbd7475bc/ovn-controller/0.log" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.601267 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.602765 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k4vvh_8729e6fd-0547-43a9-bbef-d4dcbd7475bc/ovn-controller/0.log" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.602800 4959 generic.go:334] "Generic (PLEG): container finished" podID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerID="d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119" exitCode=137 Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.602894 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-677d776894-ljrrj" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603268 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603460 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603471 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603309 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementaa19-account-delete-twnzm" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603368 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603419 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance6565-account-delete-q7tmg" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603335 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3cd4-account-delete-f4dn6" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603512 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh" event={"ID":"8729e6fd-0547-43a9-bbef-d4dcbd7475bc","Type":"ContainerDied","Data":"d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119"} Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604466 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-54b8ffdcf-7hs45" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.603273 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindercca9-account-delete-778b2" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604543 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef0b-account-delete-qnmr7" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604596 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604616 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604706 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604710 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron3e00-account-delete-qthlw" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604737 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican2a29-account-delete-fhcv4" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.604778 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.642490 4959 scope.go:117] "RemoveContainer" containerID="e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.643050 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.644750 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.671069 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.671098 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.707035 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.734061 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.751179 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4d265899-c75a-462f-9779-c573bd35a8d4" (UID: "4d265899-c75a-462f-9779-c573bd35a8d4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.762399 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30542159-188f-4508-a1b4-70285f59aef3" (UID: "30542159-188f-4508-a1b4-70285f59aef3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.766230 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data" (OuterVolumeSpecName: "config-data") pod "ff0c0088-66f1-4f85-8a34-a1f34867a37a" (UID: "ff0c0088-66f1-4f85-8a34-a1f34867a37a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.773549 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.773668 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.773724 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.777840 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.777891 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run" (OuterVolumeSpecName: "var-run") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.779930 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780031 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780083 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780246 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmdvz\" (UniqueName: \"kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780347 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780377 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780407 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780468 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltlw9\" (UniqueName: \"kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780565 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780589 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780604 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts\") pod \"9c47f87a-99bc-4191-b494-8a53e5e0298c\" (UID: \"9c47f87a-99bc-4191-b494-8a53e5e0298c\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.780670 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts\") pod \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\" (UID: \"8729e6fd-0547-43a9-bbef-d4dcbd7475bc\") " Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.782939 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785419 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785448 4959 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785464 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785479 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785493 4959 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785506 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785521 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785535 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785549 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.785562 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c47f87a-99bc-4191-b494-8a53e5e0298c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.793506 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts" (OuterVolumeSpecName: "scripts") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.801475 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz" (OuterVolumeSpecName: "kube-api-access-cmdvz") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "kube-api-access-cmdvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.812843 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.813355 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9" (OuterVolumeSpecName: "kube-api-access-ltlw9") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "kube-api-access-ltlw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.823370 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts" (OuterVolumeSpecName: "scripts") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.823505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "30542159-188f-4508-a1b4-70285f59aef3" (UID: "30542159-188f-4508-a1b4-70285f59aef3"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.828780 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data" (OuterVolumeSpecName: "config-data") pod "4d265899-c75a-462f-9779-c573bd35a8d4" (UID: "4d265899-c75a-462f-9779-c573bd35a8d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.829014 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "632a7ace-4d5d-4588-af93-e0ab25125a14" (UID: "632a7ace-4d5d-4588-af93-e0ab25125a14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.830106 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.833558 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.873844 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ec784f6-4b46-435e-af12-e899033f8348" path="/var/lib/kubelet/pods/2ec784f6-4b46-435e-af12-e899033f8348/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.875975 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" path="/var/lib/kubelet/pods/33410f10-ea8e-4ee3-be69-01f1cb9fe9ff/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.877806 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" path="/var/lib/kubelet/pods/3edab447-50e9-4fae-a95f-d5a7cb489424/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.878416 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593264f5-6188-4861-9d83-6918034eb46e" path="/var/lib/kubelet/pods/593264f5-6188-4861-9d83-6918034eb46e/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.880866 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" path="/var/lib/kubelet/pods/71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.881736 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" path="/var/lib/kubelet/pods/adc95387-65d6-40c6-82fb-e9a7f20bc514/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.882441 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" path="/var/lib/kubelet/pods/f2f9c520-72f3-482f-be82-8a625a24f3e1/volumes" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887139 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887173 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887220 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmdvz\" (UniqueName: \"kubernetes.io/projected/9c47f87a-99bc-4191-b494-8a53e5e0298c-kube-api-access-cmdvz\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887234 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d265899-c75a-462f-9779-c573bd35a8d4-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887246 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltlw9\" (UniqueName: \"kubernetes.io/projected/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-kube-api-access-ltlw9\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887255 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887263 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887292 4959 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/30542159-188f-4508-a1b4-70285f59aef3-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887302 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887311 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.887871 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data" (OuterVolumeSpecName: "config-data") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.946480 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff0c0088-66f1-4f85-8a34-a1f34867a37a" (UID: "ff0c0088-66f1-4f85-8a34-a1f34867a37a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.946797 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data" (OuterVolumeSpecName: "config-data") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:35 crc kubenswrapper[4959]: I1003 13:56:35.964678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.007466 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.007981 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff0c0088-66f1-4f85-8a34-a1f34867a37a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.008063 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.007515 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data" (OuterVolumeSpecName: "config-data") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.008135 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.032508 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.049342 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf" (OuterVolumeSpecName: "server-conf") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.053396 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.057320 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data" (OuterVolumeSpecName: "config-data") pod "632a7ace-4d5d-4588-af93-e0ab25125a14" (UID: "632a7ace-4d5d-4588-af93-e0ab25125a14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.060905 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.070412 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.081483 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data" (OuterVolumeSpecName: "config-data") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.100697 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "9ceadbac-10d3-476b-8250-5005c4391e71" (UID: "9ceadbac-10d3-476b-8250-5005c4391e71"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.101248 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.103528 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" (UID: "1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.104233 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d2a34ba-1b42-4651-93c7-8855f2398ff5" (UID: "8d2a34ba-1b42-4651-93c7-8855f2398ff5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109579 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/632a7ace-4d5d-4588-af93-e0ab25125a14-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109603 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109614 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109622 4959 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ceadbac-10d3-476b-8250-5005c4391e71-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109631 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109639 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109647 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109655 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109663 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109671 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109680 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2a34ba-1b42-4651-93c7-8855f2398ff5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.109689 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.114826 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.114858 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.114845 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f51595d7-80bf-4436-9429-9f0da918850d" (UID: "f51595d7-80bf-4436-9429-9f0da918850d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.119108 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf" (OuterVolumeSpecName: "server-conf") pod "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" (UID: "d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.124743 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.126002 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "8729e6fd-0547-43a9-bbef-d4dcbd7475bc" (UID: "8729e6fd-0547-43a9-bbef-d4dcbd7475bc"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.128805 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "ade995a8-1a92-43c8-9018-583d339f5f01" (UID: "ade995a8-1a92-43c8-9018-583d339f5f01"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.185680 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data" (OuterVolumeSpecName: "config-data") pod "9c47f87a-99bc-4191-b494-8a53e5e0298c" (UID: "9c47f87a-99bc-4191-b494-8a53e5e0298c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.210656 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.210863 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8729e6fd-0547-43a9-bbef-d4dcbd7475bc-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211016 4959 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade995a8-1a92-43c8-9018-583d339f5f01-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211102 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51595d7-80bf-4436-9429-9f0da918850d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211215 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211292 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211361 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.211438 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c47f87a-99bc-4191-b494-8a53e5e0298c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.225414 4959 scope.go:117] "RemoveContainer" containerID="79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6" Oct 03 13:56:36 crc kubenswrapper[4959]: E1003 13:56:36.231441 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6\": container with ID starting with 79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6 not found: ID does not exist" containerID="79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.231518 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6"} err="failed to get container status \"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6\": rpc error: code = NotFound desc = could not find container \"79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6\": container with ID starting with 79d8ae8d85d2299a9b7960baf0a9a65251c96f9e51d3f457609581690a1b7cd6 not found: ID does not exist" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.231565 4959 scope.go:117] "RemoveContainer" containerID="e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307" Oct 03 13:56:36 crc kubenswrapper[4959]: E1003 13:56:36.232105 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307\": container with ID starting with e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307 not found: ID does not exist" containerID="e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.232130 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307"} err="failed to get container status \"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307\": rpc error: code = NotFound desc = could not find container \"e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307\": container with ID starting with e49fa16130da74b96f7244068e9d00c42f35b34137c9629c41b2fb6ea8b03307 not found: ID does not exist" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.232145 4959 scope.go:117] "RemoveContainer" containerID="d50f3468c00aae01ce893bf689155cc435ce99dc76ce65e1b2f85525ca437119" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.284841 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.413883 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9qp4\" (UniqueName: \"kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4\") pod \"774fd066-a8d1-449a-8e33-37052dbb1627\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.413994 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data\") pod \"774fd066-a8d1-449a-8e33-37052dbb1627\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.414036 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle\") pod \"774fd066-a8d1-449a-8e33-37052dbb1627\" (UID: \"774fd066-a8d1-449a-8e33-37052dbb1627\") " Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.445379 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4" (OuterVolumeSpecName: "kube-api-access-f9qp4") pod "774fd066-a8d1-449a-8e33-37052dbb1627" (UID: "774fd066-a8d1-449a-8e33-37052dbb1627"). InnerVolumeSpecName "kube-api-access-f9qp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.459102 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "774fd066-a8d1-449a-8e33-37052dbb1627" (UID: "774fd066-a8d1-449a-8e33-37052dbb1627"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.470863 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.474099 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data" (OuterVolumeSpecName: "config-data") pod "774fd066-a8d1-449a-8e33-37052dbb1627" (UID: "774fd066-a8d1-449a-8e33-37052dbb1627"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.478083 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementaa19-account-delete-twnzm"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.485336 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.499026 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.510107 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.516328 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.516391 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/774fd066-a8d1-449a-8e33-37052dbb1627-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.516408 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9qp4\" (UniqueName: \"kubernetes.io/projected/774fd066-a8d1-449a-8e33-37052dbb1627-kube-api-access-f9qp4\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.525766 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi3cd4-account-delete-f4dn6"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.540790 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.563523 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.588470 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.601347 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0ef0b-account-delete-qnmr7"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.621406 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c47f87a-99bc-4191-b494-8a53e5e0298c","Type":"ContainerDied","Data":"008fd636ed07ff2229a285a6c6a34162f656b78166e80c2b35d0f103b465bcd6"} Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.621461 4959 scope.go:117] "RemoveContainer" containerID="ada15b2c15fe2f2dacc57422fd81cb3d234b56614971b21c30ab71da932c7b32" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.621465 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.624947 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.626002 4959 generic.go:334] "Generic (PLEG): container finished" podID="774fd066-a8d1-449a-8e33-37052dbb1627" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" exitCode=0 Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.626065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"774fd066-a8d1-449a-8e33-37052dbb1627","Type":"ContainerDied","Data":"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f"} Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.626084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"774fd066-a8d1-449a-8e33-37052dbb1627","Type":"ContainerDied","Data":"062a5da45e7aee48cc0feaa8f34c452c68d4bd35bcb03eab23fbb3077b5620f6"} Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.626129 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.628473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4vvh" event={"ID":"8729e6fd-0547-43a9-bbef-d4dcbd7475bc","Type":"ContainerDied","Data":"61e7020ede208fc0608d9a051b91870568fedaded24984255e5a778a1f2461e6"} Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.628613 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4vvh" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.639465 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron3e00-account-delete-qthlw"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.663134 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.664608 4959 scope.go:117] "RemoveContainer" containerID="13ffbebcbaac00743c439cf6bbe0144cf9261560d362f82d27b626d91213b204" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.673385 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican2a29-account-delete-fhcv4"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.685298 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.709560 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cindercca9-account-delete-778b2"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.722010 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.727005 4959 scope.go:117] "RemoveContainer" containerID="47160c83891466f0f38183ffc99e4cb435c5e8b448cf8d1325e15334ebeb44d0" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.745751 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-54b8ffdcf-7hs45"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.757421 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.765929 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance6565-account-delete-q7tmg"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.782301 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.167:8776/healthcheck\": context deadline exceeded" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.782319 4959 scope.go:117] "RemoveContainer" containerID="4ddb932943c117d08c1006b64f264db30d2dae0973fb4d5017d89ad84b3ec3a8" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.786306 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.793755 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.807811 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.817316 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: E1003 13:56:36.818318 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaa51dc6_b33c_41fc_9829_e4d20da77df3.slice/crio-90f1041e8b84ec05e8836c746e818338864fed1219609de8452c18aae574bc04\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d265899_c75a_462f_9779_c573bd35a8d4.slice/crio-caae51897f61d45af2f7174fb8c8b4b34f4a757768c222ebc005f304dfb827c6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0c0088_66f1_4f85_8a34_a1f34867a37a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2a34ba_1b42_4651_93c7_8855f2398ff5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ceadbac_10d3_476b_8250_5005c4391e71.slice/crio-46d4ef5433828742eb01485661355e76598207f1e613444f114bde9f8b93874c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8729e6fd_0547_43a9_bbef_d4dcbd7475bc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf51595d7_80bf_4436_9429_9f0da918850d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod632a7ace_4d5d_4588_af93_e0ab25125a14.slice/crio-e8bf7897e8fce36e2d1c3db77268b60ee10b6eff36ec7d80fc694ca8b60fe07b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5d4f548_e25f_4c76_a5c4_d0cf71db3ef2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade995a8_1a92_43c8_9018_583d339f5f01.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c47f87a_99bc_4191_b494_8a53e5e0298c.slice/crio-008fd636ed07ff2229a285a6c6a34162f656b78166e80c2b35d0f103b465bcd6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c47f87a_99bc_4191_b494_8a53e5e0298c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d265899_c75a_462f_9779_c573bd35a8d4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff0c0088_66f1_4f85_8a34_a1f34867a37a.slice/crio-043834d7ee8c98de79c62bb16d28a810049cad95f4a61c82e4ddbee617d92182\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade995a8_1a92_43c8_9018_583d339f5f01.slice/crio-d25d5d7ca31fbf65ba57aee081929fb858eb137519251181c4cf138741675524\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaa51dc6_b33c_41fc_9829_e4d20da77df3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf51595d7_80bf_4436_9429_9f0da918850d.slice/crio-e124d007b9a34f34f29dc16adecbd1b15d486bea47f1f2b2c79f97bf622bf1cd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30542159_188f_4508_a1b4_70285f59aef3.slice/crio-5679e93fed72dd8b41453e5689ba161f9fa1b4de849be2de8393badee0b9d51e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c9b61e3_0ba5_4abf_ac37_37e5395e2d6d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c9b61e3_0ba5_4abf_ac37_37e5395e2d6d.slice/crio-8f766212cbcf03f996ab6cfb987a50748c1c38054d97ab3ae36c0737e58417ce\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ceadbac_10d3_476b_8250_5005c4391e71.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53ba77ff_6205_4215_bb28_a45ea0b7b39f.slice\": RecentStats: unable to find data in memory cache]" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.825281 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.827635 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.827681 4959 scope.go:117] "RemoveContainer" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.832574 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.837874 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-677d776894-ljrrj"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.842723 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.858254 4959 scope.go:117] "RemoveContainer" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" Oct 03 13:56:36 crc kubenswrapper[4959]: E1003 13:56:36.862393 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f\": container with ID starting with c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f not found: ID does not exist" containerID="c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.862433 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f"} err="failed to get container status \"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f\": rpc error: code = NotFound desc = could not find container \"c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f\": container with ID starting with c8c9cf7ea739a425ce3e44d6df75ad449f701f46b923456369da74446a64bb3f not found: ID does not exist" Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.862447 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5b6c4d45db-cwzjn"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.868466 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.872783 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.876913 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.881252 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.888235 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.892338 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.896272 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.900519 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.904965 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.909534 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-k4vvh"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.920224 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.935593 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.935663 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:56:36 crc kubenswrapper[4959]: I1003 13:56:36.941550 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.682890 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-677d776894-ljrrj" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": dial tcp 10.217.0.150:9311: i/o timeout" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.683405 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-677d776894-ljrrj" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.702653 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" path="/var/lib/kubelet/pods/1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.703336 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da07093-5ea1-464a-a806-d30494ae9929" path="/var/lib/kubelet/pods/1da07093-5ea1-464a-a806-d30494ae9929/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.703850 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30542159-188f-4508-a1b4-70285f59aef3" path="/var/lib/kubelet/pods/30542159-188f-4508-a1b4-70285f59aef3/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.704895 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e788618-6d73-4794-8b04-ef41f70a94ab" path="/var/lib/kubelet/pods/3e788618-6d73-4794-8b04-ef41f70a94ab/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.705401 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" path="/var/lib/kubelet/pods/4d265899-c75a-462f-9779-c573bd35a8d4/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.705889 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ba77ff-6205-4215-bb28-a45ea0b7b39f" path="/var/lib/kubelet/pods/53ba77ff-6205-4215-bb28-a45ea0b7b39f/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.706925 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" path="/var/lib/kubelet/pods/632a7ace-4d5d-4588-af93-e0ab25125a14/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.707414 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" path="/var/lib/kubelet/pods/774fd066-a8d1-449a-8e33-37052dbb1627/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.707987 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" path="/var/lib/kubelet/pods/8729e6fd-0547-43a9-bbef-d4dcbd7475bc/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.709139 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" path="/var/lib/kubelet/pods/8d2a34ba-1b42-4651-93c7-8855f2398ff5/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.709643 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="943c9966-87bc-44fb-a19f-4ae15db1bdc4" path="/var/lib/kubelet/pods/943c9966-87bc-44fb-a19f-4ae15db1bdc4/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.710111 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" path="/var/lib/kubelet/pods/9c47f87a-99bc-4191-b494-8a53e5e0298c/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.711351 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" path="/var/lib/kubelet/pods/9ceadbac-10d3-476b-8250-5005c4391e71/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.711976 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a576d69d-a298-42d5-9c41-de82172c7a32" path="/var/lib/kubelet/pods/a576d69d-a298-42d5-9c41-de82172c7a32/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.712578 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" path="/var/lib/kubelet/pods/aaa51dc6-b33c-41fc-9829-e4d20da77df3/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.713771 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" path="/var/lib/kubelet/pods/ade995a8-1a92-43c8-9018-583d339f5f01/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.714249 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" path="/var/lib/kubelet/pods/d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.714731 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" path="/var/lib/kubelet/pods/d375128e-6fb8-4c35-9def-4bd67c206d39/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.716105 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" path="/var/lib/kubelet/pods/d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.716673 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f51595d7-80bf-4436-9429-9f0da918850d" path="/var/lib/kubelet/pods/f51595d7-80bf-4436-9429-9f0da918850d/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.717549 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f870f35a-1174-4568-9abb-b63d381483a2" path="/var/lib/kubelet/pods/f870f35a-1174-4568-9abb-b63d381483a2/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.718004 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" path="/var/lib/kubelet/pods/ff0c0088-66f1-4f85-8a34-a1f34867a37a/volumes" Oct 03 13:56:37 crc kubenswrapper[4959]: I1003 13:56:37.805418 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="593264f5-6188-4861-9d83-6918034eb46e" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.189:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:38 crc kubenswrapper[4959]: I1003 13:56:38.099450 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="30542159-188f-4508-a1b4-70285f59aef3" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.102:11211: i/o timeout" Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.379732 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.380089 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.380472 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.380506 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.380884 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.382099 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.383309 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:38 crc kubenswrapper[4959]: E1003 13:56:38.383372 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:39 crc kubenswrapper[4959]: I1003 13:56:39.948865 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:39 crc kubenswrapper[4959]: I1003 13:56:39.948865 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:41 crc kubenswrapper[4959]: E1003 13:56:41.611674 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:41 crc kubenswrapper[4959]: E1003 13:56:41.612045 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:56:57.612018992 +0000 UTC m=+1586.815362449 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.199562 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.227560 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.227892 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.228009 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.228087 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.228233 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.228355 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d99lt\" (UniqueName: \"kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.228415 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs\") pod \"84716cd9-f863-4713-bc0e-15e719ac29d2\" (UID: \"84716cd9-f863-4713-bc0e-15e719ac29d2\") " Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.258837 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.259365 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt" (OuterVolumeSpecName: "kube-api-access-d99lt") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "kube-api-access-d99lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.291741 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config" (OuterVolumeSpecName: "config") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.306370 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.307288 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.311688 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.319100 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "84716cd9-f863-4713-bc0e-15e719ac29d2" (UID: "84716cd9-f863-4713-bc0e-15e719ac29d2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330204 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330240 4959 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330250 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330259 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330268 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330277 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d99lt\" (UniqueName: \"kubernetes.io/projected/84716cd9-f863-4713-bc0e-15e719ac29d2-kube-api-access-d99lt\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.330285 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84716cd9-f863-4713-bc0e-15e719ac29d2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.710180 4959 generic.go:334] "Generic (PLEG): container finished" podID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerID="fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808" exitCode=0 Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.710251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerDied","Data":"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808"} Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.710915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76dd7cdf5-g59pr" event={"ID":"84716cd9-f863-4713-bc0e-15e719ac29d2","Type":"ContainerDied","Data":"e98eb48a73086f65135a709ca43532fe7c2309a0b2ab502d99924bf749bfb87e"} Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.710327 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76dd7cdf5-g59pr" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.710947 4959 scope.go:117] "RemoveContainer" containerID="dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.765770 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.769640 4959 scope.go:117] "RemoveContainer" containerID="fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.774000 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-76dd7cdf5-g59pr"] Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.821622 4959 scope.go:117] "RemoveContainer" containerID="dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6" Oct 03 13:56:42 crc kubenswrapper[4959]: E1003 13:56:42.822133 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6\": container with ID starting with dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6 not found: ID does not exist" containerID="dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.822172 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6"} err="failed to get container status \"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6\": rpc error: code = NotFound desc = could not find container \"dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6\": container with ID starting with dd630cf25dcf981e4fe334b2759554f93652213a9f8216670d65629ddcaaa6e6 not found: ID does not exist" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.822210 4959 scope.go:117] "RemoveContainer" containerID="fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808" Oct 03 13:56:42 crc kubenswrapper[4959]: E1003 13:56:42.822560 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808\": container with ID starting with fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808 not found: ID does not exist" containerID="fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808" Oct 03 13:56:42 crc kubenswrapper[4959]: I1003 13:56:42.822617 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808"} err="failed to get container status \"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808\": rpc error: code = NotFound desc = could not find container \"fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808\": container with ID starting with fcf0469df5cc9116674903bbac973e1124c37acc8b06c04064997458230a2808 not found: ID does not exist" Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.377839 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.378326 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.378597 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.378659 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.379121 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.380566 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.381738 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:43 crc kubenswrapper[4959]: E1003 13:56:43.381776 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:43 crc kubenswrapper[4959]: I1003 13:56:43.697699 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" path="/var/lib/kubelet/pods/84716cd9-f863-4713-bc0e-15e719ac29d2/volumes" Oct 03 13:56:45 crc kubenswrapper[4959]: I1003 13:56:45.748480 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:45 crc kubenswrapper[4959]: I1003 13:56:45.748523 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.378883 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.380355 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.381073 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.381902 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.381983 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.382491 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.385136 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:48 crc kubenswrapper[4959]: E1003 13:56:48.385255 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:50 crc kubenswrapper[4959]: I1003 13:56:50.685143 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:56:50 crc kubenswrapper[4959]: E1003 13:56:50.685692 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:56:50 crc kubenswrapper[4959]: I1003 13:56:50.757375 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:50 crc kubenswrapper[4959]: I1003 13:56:50.757379 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.379117 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.380399 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.380921 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.381014 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.382176 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.384291 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.386348 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 03 13:56:53 crc kubenswrapper[4959]: E1003 13:56:53.386430 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-2bhp5" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:56:55 crc kubenswrapper[4959]: I1003 13:56:55.767536 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:55 crc kubenswrapper[4959]: I1003 13:56:55.767542 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.709015 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2bhp5_303c0421-debb-437b-8a12-4ff587f419f3/ovs-vswitchd/0.log" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.710300 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.871882 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872034 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872047 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log" (OuterVolumeSpecName: "var-log") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872099 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872128 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872159 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4txg4\" (UniqueName: \"kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872179 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib\") pod \"303c0421-debb-437b-8a12-4ff587f419f3\" (UID: \"303c0421-debb-437b-8a12-4ff587f419f3\") " Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872281 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run" (OuterVolumeSpecName: "var-run") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872388 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib" (OuterVolumeSpecName: "var-lib") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872652 4959 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-log\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872667 4959 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872677 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.872688 4959 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/303c0421-debb-437b-8a12-4ff587f419f3-var-lib\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.873292 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts" (OuterVolumeSpecName: "scripts") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.898088 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4" (OuterVolumeSpecName: "kube-api-access-4txg4") pod "303c0421-debb-437b-8a12-4ff587f419f3" (UID: "303c0421-debb-437b-8a12-4ff587f419f3"). InnerVolumeSpecName "kube-api-access-4txg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.903399 4959 generic.go:334] "Generic (PLEG): container finished" podID="29a02381-a818-4890-9c9a-089fc14e565d" containerID="8d597ede10216948eca70fee9a0011ee11ae068e5bb066f3921145f9f96049e6" exitCode=137 Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.903484 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"8d597ede10216948eca70fee9a0011ee11ae068e5bb066f3921145f9f96049e6"} Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.905943 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2bhp5_303c0421-debb-437b-8a12-4ff587f419f3/ovs-vswitchd/0.log" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.906963 4959 generic.go:334] "Generic (PLEG): container finished" podID="303c0421-debb-437b-8a12-4ff587f419f3" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" exitCode=137 Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.906998 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerDied","Data":"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca"} Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.907021 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2bhp5" event={"ID":"303c0421-debb-437b-8a12-4ff587f419f3","Type":"ContainerDied","Data":"079309bc50bf499106e77138fdf5d99d3c5cd83c60ecdcc55627bfead97d9b11"} Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.907041 4959 scope.go:117] "RemoveContainer" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.907181 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2bhp5" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.943790 4959 scope.go:117] "RemoveContainer" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.962650 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.968940 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-2bhp5"] Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.974039 4959 scope.go:117] "RemoveContainer" containerID="354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.974740 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/303c0421-debb-437b-8a12-4ff587f419f3-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.974789 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4txg4\" (UniqueName: \"kubernetes.io/projected/303c0421-debb-437b-8a12-4ff587f419f3-kube-api-access-4txg4\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.999246 4959 scope.go:117] "RemoveContainer" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" Oct 03 13:56:56 crc kubenswrapper[4959]: E1003 13:56:56.999696 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca\": container with ID starting with ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca not found: ID does not exist" containerID="ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.999742 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca"} err="failed to get container status \"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca\": rpc error: code = NotFound desc = could not find container \"ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca\": container with ID starting with ad218069be633eb3968ac67b8b7f458db17645dd32a0b379f22854ec69056fca not found: ID does not exist" Oct 03 13:56:56 crc kubenswrapper[4959]: I1003 13:56:56.999769 4959 scope.go:117] "RemoveContainer" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" Oct 03 13:56:57 crc kubenswrapper[4959]: E1003 13:56:57.000082 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb\": container with ID starting with 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb not found: ID does not exist" containerID="68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.000126 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb"} err="failed to get container status \"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb\": rpc error: code = NotFound desc = could not find container \"68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb\": container with ID starting with 68c8b20acf55d9b03d3c5c6eabddcf449e4011794abd9a3de5e6004ec27936eb not found: ID does not exist" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.000158 4959 scope.go:117] "RemoveContainer" containerID="354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e" Oct 03 13:56:57 crc kubenswrapper[4959]: E1003 13:56:57.000615 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e\": container with ID starting with 354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e not found: ID does not exist" containerID="354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.000641 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e"} err="failed to get container status \"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e\": rpc error: code = NotFound desc = could not find container \"354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e\": container with ID starting with 354b6556006ddd4b5b11662aec1c5a9348269423fbf264a2d25cdbf30937332e not found: ID does not exist" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.627078 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:56:57 crc kubenswrapper[4959]: E1003 13:56:57.683708 4959 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Oct 03 13:56:57 crc kubenswrapper[4959]: E1003 13:56:57.683785 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data podName:71930c60-fa28-4810-98ca-6406778f917e nodeName:}" failed. No retries permitted until 2025-10-03 13:57:29.683768801 +0000 UTC m=+1618.887112218 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data") pod "barbican-api-544d968d58-6cpgj" (UID: "71930c60-fa28-4810-98ca-6406778f917e") : secret "barbican-config-data" not found Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.693027 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303c0421-debb-437b-8a12-4ff587f419f3" path="/var/lib/kubelet/pods/303c0421-debb-437b-8a12-4ff587f419f3/volumes" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.785477 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") pod \"29a02381-a818-4890-9c9a-089fc14e565d\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.785662 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache\") pod \"29a02381-a818-4890-9c9a-089fc14e565d\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.785720 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42p2c\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c\") pod \"29a02381-a818-4890-9c9a-089fc14e565d\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.785767 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock\") pod \"29a02381-a818-4890-9c9a-089fc14e565d\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.786442 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache" (OuterVolumeSpecName: "cache") pod "29a02381-a818-4890-9c9a-089fc14e565d" (UID: "29a02381-a818-4890-9c9a-089fc14e565d"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.786840 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock" (OuterVolumeSpecName: "lock") pod "29a02381-a818-4890-9c9a-089fc14e565d" (UID: "29a02381-a818-4890-9c9a-089fc14e565d"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.787032 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"29a02381-a818-4890-9c9a-089fc14e565d\" (UID: \"29a02381-a818-4890-9c9a-089fc14e565d\") " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.787784 4959 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-cache\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.787838 4959 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/29a02381-a818-4890-9c9a-089fc14e565d-lock\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.791585 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c" (OuterVolumeSpecName: "kube-api-access-42p2c") pod "29a02381-a818-4890-9c9a-089fc14e565d" (UID: "29a02381-a818-4890-9c9a-089fc14e565d"). InnerVolumeSpecName "kube-api-access-42p2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.793196 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "29a02381-a818-4890-9c9a-089fc14e565d" (UID: "29a02381-a818-4890-9c9a-089fc14e565d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.794460 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "swift") pod "29a02381-a818-4890-9c9a-089fc14e565d" (UID: "29a02381-a818-4890-9c9a-089fc14e565d"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.888410 4959 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.888446 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42p2c\" (UniqueName: \"kubernetes.io/projected/29a02381-a818-4890-9c9a-089fc14e565d-kube-api-access-42p2c\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.888475 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.903303 4959 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.925392 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.925061 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"29a02381-a818-4890-9c9a-089fc14e565d","Type":"ContainerDied","Data":"fcb92ab976eceb4c5f0f226560479341a04e4c7eb7fe937996a882532e040c0d"} Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.925534 4959 scope.go:117] "RemoveContainer" containerID="8d597ede10216948eca70fee9a0011ee11ae068e5bb066f3921145f9f96049e6" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.946965 4959 scope.go:117] "RemoveContainer" containerID="004034aa9b447a038bca1e3cffcf1b4fa3f5c22317a682340c9b6936b0225942" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.966275 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.969890 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.983429 4959 scope.go:117] "RemoveContainer" containerID="e93efd61469a9f4593c4dda8ba711b8d800d8760fc00d01dc03296a37a56265e" Oct 03 13:56:57 crc kubenswrapper[4959]: I1003 13:56:57.989720 4959 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.000893 4959 scope.go:117] "RemoveContainer" containerID="30ca3bfd6f09224ce61168200372b479678d8278ab9c2dd7cba0dd99816f2b87" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.020067 4959 scope.go:117] "RemoveContainer" containerID="a3d3d328227c680fc55d1e2c7cd597cae86b320ddc73849d22692bc5ba674e45" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.035744 4959 scope.go:117] "RemoveContainer" containerID="6000516424e24bea6505665e93fef7a122ed78ffe7c2ba42d237e867092bc2dd" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.061801 4959 scope.go:117] "RemoveContainer" containerID="23005919d7ab8893cd7eced93743eb0b32520fcd64407c9e2193608b3091452f" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.078847 4959 scope.go:117] "RemoveContainer" containerID="326cf0aa34a60cab6dd4c8c253341c34b40b93848f039d7eab65897e840f352f" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.104293 4959 scope.go:117] "RemoveContainer" containerID="27d21e8e01afbe1d304e1134184443c9d6a5ca00c5cf331aa5561d1989f0cb19" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.125610 4959 scope.go:117] "RemoveContainer" containerID="51c570be58868ca7518ace8ff746ead4f09d193a7fef31832eeba7dc57147fd9" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.148123 4959 scope.go:117] "RemoveContainer" containerID="e40663af33c9b4d1c19be468f7912534915bb3fba53c5989f5776aa52e428838" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.175004 4959 scope.go:117] "RemoveContainer" containerID="bed191a2d6ad3263372ef130bd3b54ebdd2a1a844aa9131fc1eac84370b2993e" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.193581 4959 scope.go:117] "RemoveContainer" containerID="34ae11a48814fa28d689447bfba27eedc369d31d57a2d21cecd65f5c3ab9be8a" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.213591 4959 scope.go:117] "RemoveContainer" containerID="fc9e2bfb000766ae8566eb0974e5a4ea4297a6971e4c71891ab0065f78e71574" Oct 03 13:56:58 crc kubenswrapper[4959]: I1003 13:56:58.233685 4959 scope.go:117] "RemoveContainer" containerID="0b5395c8c9b277726906ea76a40f04683a601fd733480edda15aba4179fba2b5" Oct 03 13:56:59 crc kubenswrapper[4959]: I1003 13:56:59.702157 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a02381-a818-4890-9c9a-089fc14e565d" path="/var/lib/kubelet/pods/29a02381-a818-4890-9c9a-089fc14e565d/volumes" Oct 03 13:56:59 crc kubenswrapper[4959]: I1003 13:56:59.923136 4959 scope.go:117] "RemoveContainer" containerID="9a443e77a12743b8133a947d16386d9a7160da93b531ed00b22855dabcff87e6" Oct 03 13:56:59 crc kubenswrapper[4959]: I1003 13:56:59.950394 4959 scope.go:117] "RemoveContainer" containerID="7e936f8054f962d30143ad13618e86504074fc71bae1a46de7d769eac649be3e" Oct 03 13:56:59 crc kubenswrapper[4959]: I1003 13:56:59.982244 4959 scope.go:117] "RemoveContainer" containerID="aae8036f06d6bb33643dcc5d309e687768c23fb0437205d1473dc1f70e1e4751" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.015542 4959 scope.go:117] "RemoveContainer" containerID="f8fd5ab0642adddf2c33007bb72f30687be6bd48061330c70f00edba15da7da2" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.039054 4959 scope.go:117] "RemoveContainer" containerID="1af66b255906c8775c33d1bfe9a4fbf4e95006a4472a5b2389c20102dfd422b5" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.063561 4959 scope.go:117] "RemoveContainer" containerID="1e12c25b84bbfc17840c8beb88edad3a093b4fd3bcbb1e87e14c1aaf4a2ed572" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.098947 4959 scope.go:117] "RemoveContainer" containerID="c979a79ee658ed339b2913d7a6893fbe42a93bd9add0d47862376c97d3845ac0" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.140104 4959 scope.go:117] "RemoveContainer" containerID="0854357ab58301987a5e872493993a441399ef76425c0597c5a86e12bd783410" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.777519 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:57:00 crc kubenswrapper[4959]: I1003 13:57:00.777541 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.602871 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.671041 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs\") pod \"b963e179-793e-4441-9c4d-607ab547e0ea\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.671087 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom\") pod \"b963e179-793e-4441-9c4d-607ab547e0ea\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.671972 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs" (OuterVolumeSpecName: "logs") pod "b963e179-793e-4441-9c4d-607ab547e0ea" (UID: "b963e179-793e-4441-9c4d-607ab547e0ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.672851 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.677606 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b963e179-793e-4441-9c4d-607ab547e0ea" (UID: "b963e179-793e-4441-9c4d-607ab547e0ea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.685889 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:02 crc kubenswrapper[4959]: E1003 13:57:02.686088 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772413 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom\") pod \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772479 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data\") pod \"b963e179-793e-4441-9c4d-607ab547e0ea\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772534 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs\") pod \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772563 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data\") pod \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772588 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle\") pod \"b963e179-793e-4441-9c4d-607ab547e0ea\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772626 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6c7g\" (UniqueName: \"kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g\") pod \"b963e179-793e-4441-9c4d-607ab547e0ea\" (UID: \"b963e179-793e-4441-9c4d-607ab547e0ea\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772653 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhcsd\" (UniqueName: \"kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd\") pod \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.772676 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle\") pod \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\" (UID: \"246d3dfa-dc96-4dc7-9279-bb3603218ca1\") " Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.773780 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs" (OuterVolumeSpecName: "logs") pod "246d3dfa-dc96-4dc7-9279-bb3603218ca1" (UID: "246d3dfa-dc96-4dc7-9279-bb3603218ca1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.774325 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/246d3dfa-dc96-4dc7-9279-bb3603218ca1-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.774368 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b963e179-793e-4441-9c4d-607ab547e0ea-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.774380 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.776033 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g" (OuterVolumeSpecName: "kube-api-access-x6c7g") pod "b963e179-793e-4441-9c4d-607ab547e0ea" (UID: "b963e179-793e-4441-9c4d-607ab547e0ea"). InnerVolumeSpecName "kube-api-access-x6c7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.776153 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "246d3dfa-dc96-4dc7-9279-bb3603218ca1" (UID: "246d3dfa-dc96-4dc7-9279-bb3603218ca1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.776363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd" (OuterVolumeSpecName: "kube-api-access-fhcsd") pod "246d3dfa-dc96-4dc7-9279-bb3603218ca1" (UID: "246d3dfa-dc96-4dc7-9279-bb3603218ca1"). InnerVolumeSpecName "kube-api-access-fhcsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.792584 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "246d3dfa-dc96-4dc7-9279-bb3603218ca1" (UID: "246d3dfa-dc96-4dc7-9279-bb3603218ca1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.796511 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b963e179-793e-4441-9c4d-607ab547e0ea" (UID: "b963e179-793e-4441-9c4d-607ab547e0ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.807785 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data" (OuterVolumeSpecName: "config-data") pod "b963e179-793e-4441-9c4d-607ab547e0ea" (UID: "b963e179-793e-4441-9c4d-607ab547e0ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.826417 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data" (OuterVolumeSpecName: "config-data") pod "246d3dfa-dc96-4dc7-9279-bb3603218ca1" (UID: "246d3dfa-dc96-4dc7-9279-bb3603218ca1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875479 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6c7g\" (UniqueName: \"kubernetes.io/projected/b963e179-793e-4441-9c4d-607ab547e0ea-kube-api-access-x6c7g\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875515 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhcsd\" (UniqueName: \"kubernetes.io/projected/246d3dfa-dc96-4dc7-9279-bb3603218ca1-kube-api-access-fhcsd\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875528 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875540 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875552 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875574 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/246d3dfa-dc96-4dc7-9279-bb3603218ca1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.875586 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b963e179-793e-4441-9c4d-607ab547e0ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.986727 4959 generic.go:334] "Generic (PLEG): container finished" podID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerID="3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0" exitCode=137 Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.986815 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerDied","Data":"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0"} Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.986844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" event={"ID":"246d3dfa-dc96-4dc7-9279-bb3603218ca1","Type":"ContainerDied","Data":"e1a6304ec8e0c99f5267c0f6e0fb012f9e604ea3847f5b4fdbea30d06ae23ecf"} Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.986878 4959 scope.go:117] "RemoveContainer" containerID="3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.987709 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68677546bb-dgqv8" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.989122 4959 generic.go:334] "Generic (PLEG): container finished" podID="b963e179-793e-4441-9c4d-607ab547e0ea" containerID="65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053" exitCode=137 Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.989168 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerDied","Data":"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053"} Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.989217 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b7cd665dc-9szsx" Oct 03 13:57:02 crc kubenswrapper[4959]: I1003 13:57:02.989222 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b7cd665dc-9szsx" event={"ID":"b963e179-793e-4441-9c4d-607ab547e0ea","Type":"ContainerDied","Data":"32ab7174359f57de9bc9ba592a08d7057265165d78c313d0b5ca8cbb0551bb15"} Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.024472 4959 scope.go:117] "RemoveContainer" containerID="dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.030869 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.052969 4959 scope.go:117] "RemoveContainer" containerID="3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.053924 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0\": container with ID starting with 3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0 not found: ID does not exist" containerID="3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.053969 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0"} err="failed to get container status \"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0\": rpc error: code = NotFound desc = could not find container \"3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0\": container with ID starting with 3e33e740f28d46def7f83067420f2438b6c96fac578da2bdb00518e7d0181bc0 not found: ID does not exist" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.054001 4959 scope.go:117] "RemoveContainer" containerID="dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.056771 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb\": container with ID starting with dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb not found: ID does not exist" containerID="dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.056844 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb"} err="failed to get container status \"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb\": rpc error: code = NotFound desc = could not find container \"dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb\": container with ID starting with dc5a4a616e25a1aa56c41ec8bd3f2e2181b3946d8e18443d6c9683ee16b425cb not found: ID does not exist" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.056884 4959 scope.go:117] "RemoveContainer" containerID="65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.058159 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-b7cd665dc-9szsx"] Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.065810 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.073815 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-68677546bb-dgqv8"] Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.081422 4959 scope.go:117] "RemoveContainer" containerID="07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.115446 4959 scope.go:117] "RemoveContainer" containerID="65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.116005 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053\": container with ID starting with 65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053 not found: ID does not exist" containerID="65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.116447 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053"} err="failed to get container status \"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053\": rpc error: code = NotFound desc = could not find container \"65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053\": container with ID starting with 65169c07047bbb7e121b95001179fcc066e846cf3f7e5d1f7481343ef7e62053 not found: ID does not exist" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.116518 4959 scope.go:117] "RemoveContainer" containerID="07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.117012 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7\": container with ID starting with 07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7 not found: ID does not exist" containerID="07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.117044 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7"} err="failed to get container status \"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7\": rpc error: code = NotFound desc = could not find container \"07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7\": container with ID starting with 07ce34835d3498048445d49c40d2ff59ac2183ccece12df5cf7f00d83f14eff7 not found: ID does not exist" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.475023 4959 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poddb9f7e3a-3cea-4bfd-986b-3a026d04075f"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poddb9f7e3a-3cea-4bfd-986b-3a026d04075f] : Timed out while waiting for systemd to remove kubepods-besteffort-poddb9f7e3a_3cea_4bfd_986b_3a026d04075f.slice" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.475092 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poddb9f7e3a-3cea-4bfd-986b-3a026d04075f] : unable to destroy cgroup paths for cgroup [kubepods besteffort poddb9f7e3a-3cea-4bfd-986b-3a026d04075f] : Timed out while waiting for systemd to remove kubepods-besteffort-poddb9f7e3a_3cea_4bfd_986b_3a026d04075f.slice" pod="openstack/openstack-cell1-galera-0" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.482027 4959 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod56e5d3a5-a733-402b-b393-d07706d30c48"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod56e5d3a5-a733-402b-b393-d07706d30c48] : Timed out while waiting for systemd to remove kubepods-besteffort-pod56e5d3a5_a733_402b_b393_d07706d30c48.slice" Oct 03 13:57:03 crc kubenswrapper[4959]: E1003 13:57:03.482092 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod56e5d3a5-a733-402b-b393-d07706d30c48] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod56e5d3a5-a733-402b-b393-d07706d30c48] : Timed out while waiting for systemd to remove kubepods-besteffort-pod56e5d3a5_a733_402b_b393_d07706d30c48.slice" pod="openstack/nova-cell1-novncproxy-0" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.697582 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" path="/var/lib/kubelet/pods/246d3dfa-dc96-4dc7-9279-bb3603218ca1/volumes" Oct 03 13:57:03 crc kubenswrapper[4959]: I1003 13:57:03.698673 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" path="/var/lib/kubelet/pods/b963e179-793e-4441-9c4d-607ab547e0ea/volumes" Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.000065 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.000065 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.077135 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.086334 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.097064 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.102112 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.416726 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": EOF" Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.418314 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-544d968d58-6cpgj" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.205:9311/healthcheck\": EOF" Oct 03 13:57:04 crc kubenswrapper[4959]: I1003 13:57:04.867837 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.002369 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.002464 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckzln\" (UniqueName: \"kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.002568 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.003673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.003763 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.003835 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.003881 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs\") pod \"71930c60-fa28-4810-98ca-6406778f917e\" (UID: \"71930c60-fa28-4810-98ca-6406778f917e\") " Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.004758 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs" (OuterVolumeSpecName: "logs") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.007519 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln" (OuterVolumeSpecName: "kube-api-access-ckzln") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "kube-api-access-ckzln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.019821 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.020604 4959 generic.go:334] "Generic (PLEG): container finished" podID="71930c60-fa28-4810-98ca-6406778f917e" containerID="659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879" exitCode=137 Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.020653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerDied","Data":"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879"} Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.020684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-544d968d58-6cpgj" event={"ID":"71930c60-fa28-4810-98ca-6406778f917e","Type":"ContainerDied","Data":"7857bb4799b289a4167000cc02e9f5e9f0f6f89a66c11cc29705b6cc1a5b7f19"} Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.020704 4959 scope.go:117] "RemoveContainer" containerID="659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.020830 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-544d968d58-6cpgj" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.026791 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.039582 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data" (OuterVolumeSpecName: "config-data") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.047238 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.062185 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "71930c60-fa28-4810-98ca-6406778f917e" (UID: "71930c60-fa28-4810-98ca-6406778f917e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105298 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71930c60-fa28-4810-98ca-6406778f917e-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105339 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105355 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckzln\" (UniqueName: \"kubernetes.io/projected/71930c60-fa28-4810-98ca-6406778f917e-kube-api-access-ckzln\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105366 4959 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105377 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105391 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.105402 4959 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71930c60-fa28-4810-98ca-6406778f917e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.112747 4959 scope.go:117] "RemoveContainer" containerID="bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.138498 4959 scope.go:117] "RemoveContainer" containerID="659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879" Oct 03 13:57:05 crc kubenswrapper[4959]: E1003 13:57:05.139135 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879\": container with ID starting with 659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879 not found: ID does not exist" containerID="659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.139216 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879"} err="failed to get container status \"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879\": rpc error: code = NotFound desc = could not find container \"659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879\": container with ID starting with 659c0fb3f16e6422460d5ca37e5549de7bf349544c04eae1f386cfe281268879 not found: ID does not exist" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.139253 4959 scope.go:117] "RemoveContainer" containerID="bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a" Oct 03 13:57:05 crc kubenswrapper[4959]: E1003 13:57:05.139788 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a\": container with ID starting with bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a not found: ID does not exist" containerID="bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.139841 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a"} err="failed to get container status \"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a\": rpc error: code = NotFound desc = could not find container \"bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a\": container with ID starting with bdfaf82fc17a7186ef5af09591350957449ddf3f3828f7224710e398913aa77a not found: ID does not exist" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.353225 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.357744 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-544d968d58-6cpgj"] Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.693402 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" path="/var/lib/kubelet/pods/56e5d3a5-a733-402b-b393-d07706d30c48/volumes" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.693899 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71930c60-fa28-4810-98ca-6406778f917e" path="/var/lib/kubelet/pods/71930c60-fa28-4810-98ca-6406778f917e/volumes" Oct 03 13:57:05 crc kubenswrapper[4959]: I1003 13:57:05.694566 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" path="/var/lib/kubelet/pods/db9f7e3a-3cea-4bfd-986b-3a026d04075f/volumes" Oct 03 13:57:13 crc kubenswrapper[4959]: I1003 13:57:13.685619 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:13 crc kubenswrapper[4959]: E1003 13:57:13.686802 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:24 crc kubenswrapper[4959]: I1003 13:57:24.686560 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:24 crc kubenswrapper[4959]: E1003 13:57:24.687849 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:35 crc kubenswrapper[4959]: I1003 13:57:35.686028 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:35 crc kubenswrapper[4959]: E1003 13:57:35.687333 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:47 crc kubenswrapper[4959]: I1003 13:57:47.685987 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:47 crc kubenswrapper[4959]: E1003 13:57:47.686930 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.502918 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503840 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503855 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503873 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503885 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503901 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503909 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503920 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503925 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503932 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerName="nova-scheduler-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503938 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerName="nova-scheduler-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503946 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-expirer" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503951 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-expirer" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503959 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503964 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503974 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503980 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-server" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.503992 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="rsync" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.503999 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="rsync" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504012 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504018 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504027 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="mysql-bootstrap" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504034 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="mysql-bootstrap" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504049 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504056 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504063 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="ovsdbserver-nb" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504068 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="ovsdbserver-nb" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504077 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-reaper" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504084 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-reaper" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504092 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504097 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-server" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504108 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504115 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504124 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504131 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504142 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="dnsmasq-dns" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504151 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="dnsmasq-dns" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504160 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504166 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504175 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504181 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504226 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="mysql-bootstrap" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504236 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="mysql-bootstrap" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504249 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504256 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504266 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504273 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-server" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504285 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504291 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504303 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504309 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504316 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="setup-container" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504322 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="setup-container" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504332 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504338 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504345 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504351 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504361 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504367 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504375 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504380 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504389 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504394 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504404 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504409 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504419 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504424 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504433 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504439 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504445 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504451 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504458 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504464 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504472 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504486 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504491 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504501 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504507 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504514 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504520 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504530 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerName="nova-cell1-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504535 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerName="nova-cell1-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504546 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="943c9966-87bc-44fb-a19f-4ae15db1bdc4" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504551 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="943c9966-87bc-44fb-a19f-4ae15db1bdc4" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504559 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504564 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504570 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504576 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504585 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504591 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504598 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f51595d7-80bf-4436-9429-9f0da918850d" containerName="keystone-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504603 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f51595d7-80bf-4436-9429-9f0da918850d" containerName="keystone-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504612 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504617 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504630 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="ovsdbserver-sb" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504635 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="ovsdbserver-sb" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504648 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e788618-6d73-4794-8b04-ef41f70a94ab" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504653 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e788618-6d73-4794-8b04-ef41f70a94ab" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504663 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504669 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504679 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504684 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504693 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504698 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-server" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504705 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504712 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504718 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server-init" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504724 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server-init" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504732 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504738 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504748 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="init" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504753 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="init" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504762 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504767 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504775 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504781 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504790 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504795 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504801 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504806 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504814 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-notification-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504819 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-notification-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504829 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="setup-container" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504834 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="setup-container" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504841 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504846 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504854 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a576d69d-a298-42d5-9c41-de82172c7a32" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504859 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a576d69d-a298-42d5-9c41-de82172c7a32" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504868 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504874 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504883 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="swift-recon-cron" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504888 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="swift-recon-cron" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504896 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504901 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504907 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="593264f5-6188-4861-9d83-6918034eb46e" containerName="kube-state-metrics" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504913 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="593264f5-6188-4861-9d83-6918034eb46e" containerName="kube-state-metrics" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504921 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504926 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504934 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504939 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504948 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-central-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504954 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-central-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504962 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504968 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504974 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30542159-188f-4508-a1b4-70285f59aef3" containerName="memcached" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504979 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="30542159-188f-4508-a1b4-70285f59aef3" containerName="memcached" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.504986 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.504992 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505001 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da07093-5ea1-464a-a806-d30494ae9929" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505006 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da07093-5ea1-464a-a806-d30494ae9929" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505015 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="cinder-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505020 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="cinder-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505028 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505034 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505041 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="probe" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505047 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="probe" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505054 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f870f35a-1174-4568-9abb-b63d381483a2" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505061 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f870f35a-1174-4568-9abb-b63d381483a2" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505071 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505077 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505085 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505090 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505100 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ba77ff-6205-4215-bb28-a45ea0b7b39f" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505107 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ba77ff-6205-4215-bb28-a45ea0b7b39f" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505117 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505122 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.505129 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="sg-core" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505135 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="sg-core" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505284 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-expirer" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505295 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0c0088-66f1-4f85-8a34-a1f34867a37a" containerName="nova-cell1-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505304 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="943c9966-87bc-44fb-a19f-4ae15db1bdc4" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505313 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505320 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505326 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="sg-core" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505332 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="593264f5-6188-4861-9d83-6918034eb46e" containerName="kube-state-metrics" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505341 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9b61e3-0ba5-4abf-ac37-37e5395e2d6d" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505351 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505431 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505443 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-notification-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505452 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovs-vswitchd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505460 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505469 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505476 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ba77ff-6205-4215-bb28-a45ea0b7b39f" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505483 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505490 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4bfc789-a017-43e7-ba4f-96b93514bcf4" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505498 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505506 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e788618-6d73-4794-8b04-ef41f70a94ab" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505515 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d18dc1fe-0df3-427b-8429-7fa2a6f0bf4b" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505524 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="30542159-188f-4508-a1b4-70285f59aef3" containerName="memcached" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505532 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="ceilometer-central-agent" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505539 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505546 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505551 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="33410f10-ea8e-4ee3-be69-01f1cb9fe9ff" containerName="glance-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505558 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505567 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="rsync" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505579 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="774fd066-a8d1-449a-8e33-37052dbb1627" containerName="nova-cell0-conductor-conductor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505591 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505597 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="openstack-network-exporter" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505603 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505616 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505624 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-updater" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505634 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-auditor" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505644 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505651 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c2bd3a-c4a2-4ad1-bfce-85f0194dfb1a" containerName="cinder-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505659 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505668 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505678 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505687 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5d4f548-e25f-4c76-a5c4-d0cf71db3ef2" containerName="rabbitmq" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505698 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f51595d7-80bf-4436-9429-9f0da918850d" containerName="keystone-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505706 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="84716cd9-f863-4713-bc0e-15e719ac29d2" containerName="neutron-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505714 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb01d41b-854a-4072-8b05-807edc22812f" containerName="dnsmasq-dns" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505721 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbba504-e66c-4dbe-abe3-4e69ceec1a23" containerName="ovsdbserver-sb" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505730 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8729e6fd-0547-43a9-bbef-d4dcbd7475bc" containerName="ovn-controller" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505742 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505753 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d375128e-6fb8-4c35-9def-4bd67c206d39" containerName="nova-api-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505762 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f325dd29-3b0d-4188-9ac2-e681d56e881f" containerName="proxy-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505772 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b963e179-793e-4441-9c4d-607ab547e0ea" containerName="barbican-worker-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505779 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a576d69d-a298-42d5-9c41-de82172c7a32" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505787 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f870f35a-1174-4568-9abb-b63d381483a2" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505794 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="object-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505804 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2a34ba-1b42-4651-93c7-8855f2398ff5" containerName="barbican-api-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505810 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="swift-recon-cron" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505819 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc95387-65d6-40c6-82fb-e9a7f20bc514" containerName="glance-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505825 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-reaper" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505831 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="246d3dfa-dc96-4dc7-9279-bb3603218ca1" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505843 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ceadbac-10d3-476b-8250-5005c4391e71" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505850 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c47f87a-99bc-4191-b494-8a53e5e0298c" containerName="proxy-httpd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505860 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505867 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-metadata" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505876 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da07093-5ea1-464a-a806-d30494ae9929" containerName="mariadb-account-delete" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505883 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505890 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d265899-c75a-462f-9779-c573bd35a8d4" containerName="nova-metadata-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505899 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="71930c60-fa28-4810-98ca-6406778f917e" containerName="barbican-api" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505905 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7bb6a4-7314-482c-a09a-cb25b68d7e0a" containerName="ovsdbserver-nb" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505913 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3edab447-50e9-4fae-a95f-d5a7cb489424" containerName="placement-log" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505921 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="cinder-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505928 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="account-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505936 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f9c520-72f3-482f-be82-8a625a24f3e1" containerName="probe" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505944 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="632a7ace-4d5d-4588-af93-e0ab25125a14" containerName="nova-scheduler-scheduler" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505949 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ec784f6-4b46-435e-af12-e899033f8348" containerName="barbican-keystone-listener" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505956 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e5d3a5-a733-402b-b393-d07706d30c48" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505963 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9f7e3a-3cea-4bfd-986b-3a026d04075f" containerName="galera" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505969 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a02381-a818-4890-9c9a-089fc14e565d" containerName="container-replicator" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505976 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa51dc6-b33c-41fc-9829-e4d20da77df3" containerName="barbican-worker" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505983 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="303c0421-debb-437b-8a12-4ff587f419f3" containerName="ovsdb-server" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.505991 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade995a8-1a92-43c8-9018-583d339f5f01" containerName="ovn-northd" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.506929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.521474 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.635276 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.635685 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk7jr\" (UniqueName: \"kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.635732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.685587 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:57:58 crc kubenswrapper[4959]: E1003 13:57:58.685838 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.695878 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.697508 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.706754 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.736804 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.736880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.736939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk7jr\" (UniqueName: \"kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.737366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.737433 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.760067 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk7jr\" (UniqueName: \"kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr\") pod \"redhat-operators-9vk2l\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.835693 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.838404 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.838460 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzn7x\" (UniqueName: \"kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.838663 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.940177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.940752 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.940786 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzn7x\" (UniqueName: \"kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.941171 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.941228 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:58 crc kubenswrapper[4959]: I1003 13:57:58.958791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzn7x\" (UniqueName: \"kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x\") pod \"community-operators-hb4g9\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.012740 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.120820 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.541634 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.591424 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerID="af2d9afd248650b843e374014984262f9ac61780712e12ce47dd64689d803b97" exitCode=0 Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.591488 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerDied","Data":"af2d9afd248650b843e374014984262f9ac61780712e12ce47dd64689d803b97"} Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.591513 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerStarted","Data":"ca6e98d68129d9d21f63ef14909e95d020930f040dd636175920b692a00b6174"} Oct 03 13:57:59 crc kubenswrapper[4959]: I1003 13:57:59.594152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerStarted","Data":"59abd208239a3221f018398df801fea3497eac3bcdc0763955825ca0f7b6814d"} Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.571167 4959 scope.go:117] "RemoveContainer" containerID="ce93e24bd637cc785e98cb368132c035f56a137124dfa56c0e7cd03b1a996262" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.603493 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerID="8b2f0546f1c2ad9a21880903ba357a46ab5f6d954bbd1d0abd27d534c93b1f2b" exitCode=0 Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.603547 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerDied","Data":"8b2f0546f1c2ad9a21880903ba357a46ab5f6d954bbd1d0abd27d534c93b1f2b"} Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.619421 4959 scope.go:117] "RemoveContainer" containerID="87357d5f6f8004473f56f3dd117b5d6a1b35ed9b5d28836c98184b8c60d36601" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.660903 4959 scope.go:117] "RemoveContainer" containerID="1ded0051ef4f89932251feaa8d78562d05eed4ad7db125f2a1dfac0af98a16e6" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.696909 4959 scope.go:117] "RemoveContainer" containerID="79071188a4f44f92ce2ed170a1b1c0fd2bafcd83545facaa85075a5cda7b863a" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.744492 4959 scope.go:117] "RemoveContainer" containerID="0605bc3c6b468664f6897c5503b23d22fcb8a1132c5793fe99fb04a352189084" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.800749 4959 scope.go:117] "RemoveContainer" containerID="fa4db0133a9fd539ffb0778a58527848be291d332cab66fa55c6d83908ac5d53" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.837460 4959 scope.go:117] "RemoveContainer" containerID="3dc40507c120959d206031c3611237f52e593b9449af8cabe30688abc704507f" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.868630 4959 scope.go:117] "RemoveContainer" containerID="a0e8ebea4f33a6079517be7adede19a99d3c556ce12baa9b22ddbebb6d921cd7" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.897467 4959 scope.go:117] "RemoveContainer" containerID="64c14a000a4fd84bbb5d50a1259b64583f506161d19b324050350a8e3f0d3acf" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.919636 4959 scope.go:117] "RemoveContainer" containerID="c2ef1d50c81020ba55c252cdc1ea93b3c7f0bf54d6a6ff2e7051d40381b130cf" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.938633 4959 scope.go:117] "RemoveContainer" containerID="998d4944dee5b32d1a16f943f653f08e894041495c582824790d2bdc1559922a" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.955907 4959 scope.go:117] "RemoveContainer" containerID="b198669f36a1928334af1cd79e08fedc8aebaee6e9b73477d066ad5c78372a5a" Oct 03 13:58:00 crc kubenswrapper[4959]: I1003 13:58:00.978111 4959 scope.go:117] "RemoveContainer" containerID="05c8d7b052abc4c1170e3e6ba142cf474e73af12bf8462a593d2a4d55762f344" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.006260 4959 scope.go:117] "RemoveContainer" containerID="73f4e64df58083370f19d3a84fd7e6d994163b876b512f98d2ee1e62e5754b19" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.024077 4959 scope.go:117] "RemoveContainer" containerID="a7bafb2fc033fe6372313d85d5173772ae3d7f9d3ac72d25597f8bd36ea8a953" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.060965 4959 scope.go:117] "RemoveContainer" containerID="956042f00860508daaea28b26ecf1fd459cedf0fd52cb55d9ba130ca282f8dbc" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.092695 4959 scope.go:117] "RemoveContainer" containerID="4d69f154f3479983536c2706de9142a8fb43833ce35669b457243f2cc542426c" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.110317 4959 scope.go:117] "RemoveContainer" containerID="6f930a2e79f58fb03314be9aa1247f4c2514347820d5bc31d5b2e66c097d5a69" Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.616755 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerID="f6e0f92dd023a96aa0e770f145d0d08e9a94ac5db111f854dbeaea4279d54cfc" exitCode=0 Oct 03 13:58:01 crc kubenswrapper[4959]: I1003 13:58:01.616835 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerDied","Data":"f6e0f92dd023a96aa0e770f145d0d08e9a94ac5db111f854dbeaea4279d54cfc"} Oct 03 13:58:02 crc kubenswrapper[4959]: I1003 13:58:02.635797 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerStarted","Data":"32231d959a4b38710e62fa14b02740550ccf936a5787054efd64f2fdd82a7f8c"} Oct 03 13:58:02 crc kubenswrapper[4959]: I1003 13:58:02.640011 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerStarted","Data":"ba636ad2b82f38c0ddd2673f2616fc011154ca6d693abf56b32e83f4d0d4fefd"} Oct 03 13:58:02 crc kubenswrapper[4959]: I1003 13:58:02.663559 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9vk2l" podStartSLOduration=1.9057060510000001 podStartE2EDuration="4.663542114s" podCreationTimestamp="2025-10-03 13:57:58 +0000 UTC" firstStartedPulling="2025-10-03 13:57:59.594205242 +0000 UTC m=+1648.797548649" lastFinishedPulling="2025-10-03 13:58:02.352041295 +0000 UTC m=+1651.555384712" observedRunningTime="2025-10-03 13:58:02.659370622 +0000 UTC m=+1651.862714059" watchObservedRunningTime="2025-10-03 13:58:02.663542114 +0000 UTC m=+1651.866885531" Oct 03 13:58:03 crc kubenswrapper[4959]: I1003 13:58:03.649184 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerID="ba636ad2b82f38c0ddd2673f2616fc011154ca6d693abf56b32e83f4d0d4fefd" exitCode=0 Oct 03 13:58:03 crc kubenswrapper[4959]: I1003 13:58:03.649332 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerDied","Data":"ba636ad2b82f38c0ddd2673f2616fc011154ca6d693abf56b32e83f4d0d4fefd"} Oct 03 13:58:04 crc kubenswrapper[4959]: I1003 13:58:04.660351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerStarted","Data":"4f2c73227f115f324930fe2f33686fffc6daf6600e33bce79468573cc28b939f"} Oct 03 13:58:04 crc kubenswrapper[4959]: I1003 13:58:04.683101 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hb4g9" podStartSLOduration=3.235799816 podStartE2EDuration="6.683082488s" podCreationTimestamp="2025-10-03 13:57:58 +0000 UTC" firstStartedPulling="2025-10-03 13:58:00.619623115 +0000 UTC m=+1649.822966552" lastFinishedPulling="2025-10-03 13:58:04.066905767 +0000 UTC m=+1653.270249224" observedRunningTime="2025-10-03 13:58:04.679264495 +0000 UTC m=+1653.882607922" watchObservedRunningTime="2025-10-03 13:58:04.683082488 +0000 UTC m=+1653.886425915" Oct 03 13:58:08 crc kubenswrapper[4959]: I1003 13:58:08.836411 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:08 crc kubenswrapper[4959]: I1003 13:58:08.836935 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:08 crc kubenswrapper[4959]: I1003 13:58:08.886387 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:09 crc kubenswrapper[4959]: I1003 13:58:09.013718 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:09 crc kubenswrapper[4959]: I1003 13:58:09.014011 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:09 crc kubenswrapper[4959]: I1003 13:58:09.076978 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:09 crc kubenswrapper[4959]: I1003 13:58:09.766888 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:09 crc kubenswrapper[4959]: I1003 13:58:09.771229 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:10 crc kubenswrapper[4959]: I1003 13:58:10.321062 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:58:11 crc kubenswrapper[4959]: I1003 13:58:11.730693 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hb4g9" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="registry-server" containerID="cri-o://4f2c73227f115f324930fe2f33686fffc6daf6600e33bce79468573cc28b939f" gracePeriod=2 Oct 03 13:58:12 crc kubenswrapper[4959]: I1003 13:58:12.119578 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:58:12 crc kubenswrapper[4959]: I1003 13:58:12.119866 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9vk2l" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="registry-server" containerID="cri-o://32231d959a4b38710e62fa14b02740550ccf936a5787054efd64f2fdd82a7f8c" gracePeriod=2 Oct 03 13:58:12 crc kubenswrapper[4959]: I1003 13:58:12.685724 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:58:12 crc kubenswrapper[4959]: E1003 13:58:12.686020 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:58:12 crc kubenswrapper[4959]: I1003 13:58:12.748609 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerID="4f2c73227f115f324930fe2f33686fffc6daf6600e33bce79468573cc28b939f" exitCode=0 Oct 03 13:58:12 crc kubenswrapper[4959]: I1003 13:58:12.748654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerDied","Data":"4f2c73227f115f324930fe2f33686fffc6daf6600e33bce79468573cc28b939f"} Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.300983 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.346901 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities\") pod \"3a66b356-3b40-4485-8050-a6a298ac2a21\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.346965 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzn7x\" (UniqueName: \"kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x\") pod \"3a66b356-3b40-4485-8050-a6a298ac2a21\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.347015 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content\") pod \"3a66b356-3b40-4485-8050-a6a298ac2a21\" (UID: \"3a66b356-3b40-4485-8050-a6a298ac2a21\") " Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.349861 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities" (OuterVolumeSpecName: "utilities") pod "3a66b356-3b40-4485-8050-a6a298ac2a21" (UID: "3a66b356-3b40-4485-8050-a6a298ac2a21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.356229 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x" (OuterVolumeSpecName: "kube-api-access-qzn7x") pod "3a66b356-3b40-4485-8050-a6a298ac2a21" (UID: "3a66b356-3b40-4485-8050-a6a298ac2a21"). InnerVolumeSpecName "kube-api-access-qzn7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.448045 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.448083 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzn7x\" (UniqueName: \"kubernetes.io/projected/3a66b356-3b40-4485-8050-a6a298ac2a21-kube-api-access-qzn7x\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.762446 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerID="32231d959a4b38710e62fa14b02740550ccf936a5787054efd64f2fdd82a7f8c" exitCode=0 Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.762520 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerDied","Data":"32231d959a4b38710e62fa14b02740550ccf936a5787054efd64f2fdd82a7f8c"} Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.764852 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb4g9" event={"ID":"3a66b356-3b40-4485-8050-a6a298ac2a21","Type":"ContainerDied","Data":"59abd208239a3221f018398df801fea3497eac3bcdc0763955825ca0f7b6814d"} Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.764895 4959 scope.go:117] "RemoveContainer" containerID="4f2c73227f115f324930fe2f33686fffc6daf6600e33bce79468573cc28b939f" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.764942 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb4g9" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.791554 4959 scope.go:117] "RemoveContainer" containerID="ba636ad2b82f38c0ddd2673f2616fc011154ca6d693abf56b32e83f4d0d4fefd" Oct 03 13:58:13 crc kubenswrapper[4959]: I1003 13:58:13.819058 4959 scope.go:117] "RemoveContainer" containerID="8b2f0546f1c2ad9a21880903ba357a46ab5f6d954bbd1d0abd27d534c93b1f2b" Oct 03 13:58:14 crc kubenswrapper[4959]: I1003 13:58:14.938426 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a66b356-3b40-4485-8050-a6a298ac2a21" (UID: "3a66b356-3b40-4485-8050-a6a298ac2a21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:58:14 crc kubenswrapper[4959]: I1003 13:58:14.971035 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a66b356-3b40-4485-8050-a6a298ac2a21-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.007912 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.012165 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hb4g9"] Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.060385 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.173695 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content\") pod \"2c10d567-ce1c-455e-a439-a219bb8eb256\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.173767 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk7jr\" (UniqueName: \"kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr\") pod \"2c10d567-ce1c-455e-a439-a219bb8eb256\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.173823 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities\") pod \"2c10d567-ce1c-455e-a439-a219bb8eb256\" (UID: \"2c10d567-ce1c-455e-a439-a219bb8eb256\") " Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.174750 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities" (OuterVolumeSpecName: "utilities") pod "2c10d567-ce1c-455e-a439-a219bb8eb256" (UID: "2c10d567-ce1c-455e-a439-a219bb8eb256"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.178733 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr" (OuterVolumeSpecName: "kube-api-access-rk7jr") pod "2c10d567-ce1c-455e-a439-a219bb8eb256" (UID: "2c10d567-ce1c-455e-a439-a219bb8eb256"). InnerVolumeSpecName "kube-api-access-rk7jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.263027 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c10d567-ce1c-455e-a439-a219bb8eb256" (UID: "2c10d567-ce1c-455e-a439-a219bb8eb256"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.276001 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.276050 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk7jr\" (UniqueName: \"kubernetes.io/projected/2c10d567-ce1c-455e-a439-a219bb8eb256-kube-api-access-rk7jr\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.276074 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c10d567-ce1c-455e-a439-a219bb8eb256-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.700505 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" path="/var/lib/kubelet/pods/3a66b356-3b40-4485-8050-a6a298ac2a21/volumes" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.791868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vk2l" event={"ID":"2c10d567-ce1c-455e-a439-a219bb8eb256","Type":"ContainerDied","Data":"ca6e98d68129d9d21f63ef14909e95d020930f040dd636175920b692a00b6174"} Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.791950 4959 scope.go:117] "RemoveContainer" containerID="32231d959a4b38710e62fa14b02740550ccf936a5787054efd64f2fdd82a7f8c" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.792114 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vk2l" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.814429 4959 scope.go:117] "RemoveContainer" containerID="f6e0f92dd023a96aa0e770f145d0d08e9a94ac5db111f854dbeaea4279d54cfc" Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.833899 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.846177 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9vk2l"] Oct 03 13:58:15 crc kubenswrapper[4959]: I1003 13:58:15.856959 4959 scope.go:117] "RemoveContainer" containerID="af2d9afd248650b843e374014984262f9ac61780712e12ce47dd64689d803b97" Oct 03 13:58:17 crc kubenswrapper[4959]: I1003 13:58:17.696042 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" path="/var/lib/kubelet/pods/2c10d567-ce1c-455e-a439-a219bb8eb256/volumes" Oct 03 13:58:23 crc kubenswrapper[4959]: I1003 13:58:23.685906 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:58:23 crc kubenswrapper[4959]: E1003 13:58:23.688612 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:58:34 crc kubenswrapper[4959]: I1003 13:58:34.685704 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:58:34 crc kubenswrapper[4959]: E1003 13:58:34.686645 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:58:47 crc kubenswrapper[4959]: I1003 13:58:47.685407 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:58:47 crc kubenswrapper[4959]: E1003 13:58:47.686593 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:58:58 crc kubenswrapper[4959]: I1003 13:58:58.685551 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:58:58 crc kubenswrapper[4959]: E1003 13:58:58.686383 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.443845 4959 scope.go:117] "RemoveContainer" containerID="266db7a9c0bcfa5ef155b35ab4e9ef87e78ba77d6daace95769021b26bed7452" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.509317 4959 scope.go:117] "RemoveContainer" containerID="06d4a6beaf8c7baed60c86f5ef66a97b9dcd29bdf45ef077eba2ec59e7cf24e6" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.527596 4959 scope.go:117] "RemoveContainer" containerID="cfae109f6f3042bf10c7d8abaf5600ce7fac83b01ce2f4be11acd687b892aff1" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.563428 4959 scope.go:117] "RemoveContainer" containerID="c0885574e428c640960fc2f2df84effd904ebceef0a6385f1fcac690331dee73" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.617524 4959 scope.go:117] "RemoveContainer" containerID="bff1a6e521163ac3cf38cf50dfacc21bc264522ed18b6fa7a5c53f6262dbc566" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.651731 4959 scope.go:117] "RemoveContainer" containerID="1f8f42d263cd44bb7f3d741af35c4900133542e621536ce527da52ab82dbee5c" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.675279 4959 scope.go:117] "RemoveContainer" containerID="023f6f41efd8b70bb5c4bc20ca5e840eaa4f8d065b77a0f728e8d0ecf836a688" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.695553 4959 scope.go:117] "RemoveContainer" containerID="48413a96dbd55c4eac4e40c0f894a4fe721bcad741ce542a08321841cf77afd6" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.715891 4959 scope.go:117] "RemoveContainer" containerID="95dbbc2b264de16726fd5cf962dcf5b2ea23cce70901b014633e800857d2e6ac" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.735384 4959 scope.go:117] "RemoveContainer" containerID="e2a2488abba43c5130cc5e66793ff02a224a83cad1a37062cb627eb282992260" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.753911 4959 scope.go:117] "RemoveContainer" containerID="a88bf5ca4e0c7d6091ecce6d3abd4e46c1d584111080db1a1653bc5d9ffebe87" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.782977 4959 scope.go:117] "RemoveContainer" containerID="d5033b93560574e923577b90dc0e6761b4304b82f8fedaf8d5ae2041e3f74a66" Oct 03 13:59:01 crc kubenswrapper[4959]: I1003 13:59:01.800137 4959 scope.go:117] "RemoveContainer" containerID="617f60df069f8b127236c30ef29dbcf5ad86abf5cef1969b4f6275e6d0b31ca5" Oct 03 13:59:11 crc kubenswrapper[4959]: I1003 13:59:11.698273 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:59:11 crc kubenswrapper[4959]: E1003 13:59:11.699456 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:59:23 crc kubenswrapper[4959]: I1003 13:59:23.685930 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:59:23 crc kubenswrapper[4959]: E1003 13:59:23.686690 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:59:34 crc kubenswrapper[4959]: I1003 13:59:34.686364 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:59:34 crc kubenswrapper[4959]: E1003 13:59:34.687022 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 13:59:48 crc kubenswrapper[4959]: I1003 13:59:48.685267 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 13:59:48 crc kubenswrapper[4959]: E1003 13:59:48.686101 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.149860 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5"] Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150754 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150770 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150804 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150813 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150829 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150837 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150846 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150856 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150874 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150881 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4959]: E1003 14:00:00.150890 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.150897 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.151059 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c10d567-ce1c-455e-a439-a219bb8eb256" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.151095 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a66b356-3b40-4485-8050-a6a298ac2a21" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.151578 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.155369 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.155501 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.159272 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5"] Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.306535 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.307129 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.307250 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srvlc\" (UniqueName: \"kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.408859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.408939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.408964 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srvlc\" (UniqueName: \"kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.410564 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.423666 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.429985 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srvlc\" (UniqueName: \"kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc\") pod \"collect-profiles-29325000-b48l5\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.471890 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:00 crc kubenswrapper[4959]: I1003 14:00:00.934719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5"] Oct 03 14:00:01 crc kubenswrapper[4959]: I1003 14:00:01.691305 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:00:01 crc kubenswrapper[4959]: E1003 14:00:01.691812 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:00:01 crc kubenswrapper[4959]: I1003 14:00:01.826608 4959 generic.go:334] "Generic (PLEG): container finished" podID="a1910e14-a1f9-4c81-a929-7336bd920153" containerID="807e04c8d7efbb61d0e1d815bc66ec70315642f9ea107f603e0a535ab5061989" exitCode=0 Oct 03 14:00:01 crc kubenswrapper[4959]: I1003 14:00:01.826649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" event={"ID":"a1910e14-a1f9-4c81-a929-7336bd920153","Type":"ContainerDied","Data":"807e04c8d7efbb61d0e1d815bc66ec70315642f9ea107f603e0a535ab5061989"} Oct 03 14:00:01 crc kubenswrapper[4959]: I1003 14:00:01.826678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" event={"ID":"a1910e14-a1f9-4c81-a929-7336bd920153","Type":"ContainerStarted","Data":"d2bcc380291acbc6c1087ee36ad0ca503406abdc0f1c242832ea1f02ff325258"} Oct 03 14:00:01 crc kubenswrapper[4959]: I1003 14:00:01.993920 4959 scope.go:117] "RemoveContainer" containerID="8ad815a404d34e7daeb62fd5d97fc17482809b2f82c91a7f5724b4b1e3321266" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.035401 4959 scope.go:117] "RemoveContainer" containerID="a245794e4104503620a1e62233334902247449b3325e793e6c53eee6dfed9e0a" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.062414 4959 scope.go:117] "RemoveContainer" containerID="5c237f828a85840621666072b2ca9c1f967a6c52a22efbea5023d5b7c68c4fbc" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.116019 4959 scope.go:117] "RemoveContainer" containerID="0f239875c91ee1dd657942f1b4a5bee065b5388e0803e40d082ef09005a99e55" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.133797 4959 scope.go:117] "RemoveContainer" containerID="b304c4a31b60f0e9c502c5bfcdf56c0274ea52bcd6515a31c6e5973c7f3cb733" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.153515 4959 scope.go:117] "RemoveContainer" containerID="a7bbb7b660035d1128449ae672bafdccb24dae017043de5c764061910cfda76f" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.181806 4959 scope.go:117] "RemoveContainer" containerID="a7b0ceeff61941890b21a6f7d2ca35938e4cdfd15247925f0209a5119fb4f1a1" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.200229 4959 scope.go:117] "RemoveContainer" containerID="f2500fe9d0158d0ab935afc5fe14db033c74992aa7ae26ecba958ae3596c57f9" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.225767 4959 scope.go:117] "RemoveContainer" containerID="ee16252e1bbddaf8d1e22642807ff2d9e9259f29027e5a4fff2c75e96b01eb37" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.244655 4959 scope.go:117] "RemoveContainer" containerID="d8be0657675a6b800321dc24a5acf47e262aa26ce6657e72b068da8f72f7bdef" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.280091 4959 scope.go:117] "RemoveContainer" containerID="6b617a3bf18abe83bfc72274e793ce4a300223427e3b278bb1ccc7c7bec09bcc" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.309802 4959 scope.go:117] "RemoveContainer" containerID="301f13bc924830f186afba1bebfa0f509472a654adc102a985467ad83327e434" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.343886 4959 scope.go:117] "RemoveContainer" containerID="9bc9b789b1e410f14438dd275ab4df1b4886b1a5b73a1ba9e3ec567e82067001" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.368781 4959 scope.go:117] "RemoveContainer" containerID="cc4b4c055039a16bb8698b1f062f4ab7a7185cd1b62caa23f240835b1d1163dd" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.401476 4959 scope.go:117] "RemoveContainer" containerID="87681968bf3b43d810f5c797f220cf83f293f5af560789c5caf22b22226ad5c3" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.420722 4959 scope.go:117] "RemoveContainer" containerID="8c164c5247a2631e33c76e1ec7e46302a319f9bf2b75393b0785b327b5b1ed77" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.455073 4959 scope.go:117] "RemoveContainer" containerID="342111745eadc22d2c2eafdf7a166b38b56b789123de46efd4bcb84e87205c69" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.478046 4959 scope.go:117] "RemoveContainer" containerID="a7c1d4fb80394150b035616306b0249611184085fee535c8c9ed09e63d2e3572" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.505017 4959 scope.go:117] "RemoveContainer" containerID="b310fb69f996618095e2ad2c4d0857f61089302454ce26983857cb17725a0561" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.523726 4959 scope.go:117] "RemoveContainer" containerID="1687cfe1fb627021244443cd8feaa5d63671fc0f72f8a818a5dbd7ec7dc6d2a7" Oct 03 14:00:02 crc kubenswrapper[4959]: I1003 14:00:02.543724 4959 scope.go:117] "RemoveContainer" containerID="fc9d972ac9b2f06f0e0356aa259d27ac5d465cfc53bdfcc611986b0718292921" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.102159 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.253432 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume\") pod \"a1910e14-a1f9-4c81-a929-7336bd920153\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.254174 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume" (OuterVolumeSpecName: "config-volume") pod "a1910e14-a1f9-4c81-a929-7336bd920153" (UID: "a1910e14-a1f9-4c81-a929-7336bd920153"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.254341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srvlc\" (UniqueName: \"kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc\") pod \"a1910e14-a1f9-4c81-a929-7336bd920153\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.254535 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume\") pod \"a1910e14-a1f9-4c81-a929-7336bd920153\" (UID: \"a1910e14-a1f9-4c81-a929-7336bd920153\") " Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.254919 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a1910e14-a1f9-4c81-a929-7336bd920153-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.259788 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a1910e14-a1f9-4c81-a929-7336bd920153" (UID: "a1910e14-a1f9-4c81-a929-7336bd920153"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.264058 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc" (OuterVolumeSpecName: "kube-api-access-srvlc") pod "a1910e14-a1f9-4c81-a929-7336bd920153" (UID: "a1910e14-a1f9-4c81-a929-7336bd920153"). InnerVolumeSpecName "kube-api-access-srvlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.356378 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a1910e14-a1f9-4c81-a929-7336bd920153-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.356442 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srvlc\" (UniqueName: \"kubernetes.io/projected/a1910e14-a1f9-4c81-a929-7336bd920153-kube-api-access-srvlc\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.842986 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" event={"ID":"a1910e14-a1f9-4c81-a929-7336bd920153","Type":"ContainerDied","Data":"d2bcc380291acbc6c1087ee36ad0ca503406abdc0f1c242832ea1f02ff325258"} Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.843066 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2bcc380291acbc6c1087ee36ad0ca503406abdc0f1c242832ea1f02ff325258" Oct 03 14:00:03 crc kubenswrapper[4959]: I1003 14:00:03.843026 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5" Oct 03 14:00:14 crc kubenswrapper[4959]: I1003 14:00:14.686127 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:00:14 crc kubenswrapper[4959]: E1003 14:00:14.687134 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:00:29 crc kubenswrapper[4959]: I1003 14:00:29.684811 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:00:29 crc kubenswrapper[4959]: E1003 14:00:29.685718 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:00:43 crc kubenswrapper[4959]: I1003 14:00:43.685422 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:00:43 crc kubenswrapper[4959]: E1003 14:00:43.685848 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:00:55 crc kubenswrapper[4959]: I1003 14:00:55.687650 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:00:55 crc kubenswrapper[4959]: E1003 14:00:55.688909 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:01:02 crc kubenswrapper[4959]: I1003 14:01:02.787166 4959 scope.go:117] "RemoveContainer" containerID="286c3a4d76e172ce7cfa4c2cf6b0f1d71271dbefac5d465aa78b0eb9555090f9" Oct 03 14:01:02 crc kubenswrapper[4959]: I1003 14:01:02.835312 4959 scope.go:117] "RemoveContainer" containerID="f9a14d3f342cd387cfde7a61857031e7aba2a3acb904f84b2b45d694964778f3" Oct 03 14:01:02 crc kubenswrapper[4959]: I1003 14:01:02.871337 4959 scope.go:117] "RemoveContainer" containerID="35e753ac6de1e60008c5fef25525ceb59336b6d2eaf14a538f17a9e256f1f75b" Oct 03 14:01:02 crc kubenswrapper[4959]: I1003 14:01:02.890419 4959 scope.go:117] "RemoveContainer" containerID="b54e17dcdfb09d6aeb29e71dcf080822e14a3bd78cc63a8d522dcef2f390bb8d" Oct 03 14:01:02 crc kubenswrapper[4959]: I1003 14:01:02.967022 4959 scope.go:117] "RemoveContainer" containerID="0aca4e4e5b11a2f5ad93029ec98a5edd017bc0bff5d499883f3af4332959c29e" Oct 03 14:01:03 crc kubenswrapper[4959]: I1003 14:01:03.010980 4959 scope.go:117] "RemoveContainer" containerID="e0af0d6c5977be30d5b286218011035d1584aa803d27be60b4fa6412d14d8c7d" Oct 03 14:01:03 crc kubenswrapper[4959]: I1003 14:01:03.058707 4959 scope.go:117] "RemoveContainer" containerID="437ecf2e1e223998fb6b040d3dc342771553ad152a649a307ba005f342deed15" Oct 03 14:01:10 crc kubenswrapper[4959]: I1003 14:01:10.685816 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:01:11 crc kubenswrapper[4959]: I1003 14:01:11.482311 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1"} Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.204216 4959 scope.go:117] "RemoveContainer" containerID="cfb72705ebccca6e3af246841cf201b6036478e1d451c879eb40ab8cf65d5799" Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.243249 4959 scope.go:117] "RemoveContainer" containerID="72a3e7929b0adc554a8bfd78df9f7708f0538f33a7bbf4518d4a306cb7e072a8" Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.270351 4959 scope.go:117] "RemoveContainer" containerID="9f1517df64486c3bcc421232da8688f5478dc78c9cfeec76b3735eadbe8fbb25" Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.285363 4959 scope.go:117] "RemoveContainer" containerID="bd8aaedeb008acc8fdb85e7ff1a53c8c9fefc46446b823e52046fd7ccab71695" Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.302909 4959 scope.go:117] "RemoveContainer" containerID="8c342ba7a12a3d1b866f17f84e61cfc1269c690927daba2bc2a8e0c10ed68b3c" Oct 03 14:02:03 crc kubenswrapper[4959]: I1003 14:02:03.342715 4959 scope.go:117] "RemoveContainer" containerID="1456ab292b85bd9f4fc93550fd3ab07755616533cee7b5fbca04110a5065766a" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.416514 4959 scope.go:117] "RemoveContainer" containerID="55c2efbebd0db927e2a324867a441f257d17129410fe8d5306251eed8adffcef" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.452000 4959 scope.go:117] "RemoveContainer" containerID="41f0dbd820ad59e9479c3ed201b87dee36f9fad0fd1f2a3b8f965b066cdb290a" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.478517 4959 scope.go:117] "RemoveContainer" containerID="6a87988298ce622eae57f154f644d7de59597c62074a0c30695ef873834d9fc0" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.518072 4959 scope.go:117] "RemoveContainer" containerID="13f35655ec5a85e5aa845c09f3f121c12edd40f9d1186f6c00e0a06cb62b25d7" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.538913 4959 scope.go:117] "RemoveContainer" containerID="e9a759ce3223705acf4e1e483d5369ad492fba545b667e79de858b249d48d6d2" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.560281 4959 scope.go:117] "RemoveContainer" containerID="a8c979b1746ba8dacc24c61273c411303183dd2bb1f9d4920d65bb406dad43c5" Oct 03 14:03:03 crc kubenswrapper[4959]: I1003 14:03:03.593910 4959 scope.go:117] "RemoveContainer" containerID="65ae3048e36c4052ae2f42bf6462f21a9255680d29afaf9afdfbca6644e45e3c" Oct 03 14:03:36 crc kubenswrapper[4959]: I1003 14:03:36.044872 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:03:36 crc kubenswrapper[4959]: I1003 14:03:36.045579 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:04:06 crc kubenswrapper[4959]: I1003 14:04:06.043978 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:04:06 crc kubenswrapper[4959]: I1003 14:04:06.045332 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.044726 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.045740 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.045835 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.047334 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.047412 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1" gracePeriod=600 Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.265398 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1" exitCode=0 Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.265445 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1"} Oct 03 14:04:36 crc kubenswrapper[4959]: I1003 14:04:36.265487 4959 scope.go:117] "RemoveContainer" containerID="1e0b76aea94d40ab5ac6d7a88587493b6a9220d1c17aa4d7113cf734bf750c4c" Oct 03 14:04:37 crc kubenswrapper[4959]: I1003 14:04:37.276282 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a"} Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.349273 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:44 crc kubenswrapper[4959]: E1003 14:05:44.349951 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1910e14-a1f9-4c81-a929-7336bd920153" containerName="collect-profiles" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.349963 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1910e14-a1f9-4c81-a929-7336bd920153" containerName="collect-profiles" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.350110 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1910e14-a1f9-4c81-a929-7336bd920153" containerName="collect-profiles" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.351037 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.384903 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.402728 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.402836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.403123 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzps5\" (UniqueName: \"kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.504319 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.504408 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzps5\" (UniqueName: \"kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.504505 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.504979 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.505001 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.531028 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzps5\" (UniqueName: \"kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5\") pod \"redhat-marketplace-j7cr7\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:44 crc kubenswrapper[4959]: I1003 14:05:44.707898 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:45 crc kubenswrapper[4959]: I1003 14:05:45.108891 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:45 crc kubenswrapper[4959]: I1003 14:05:45.822474 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerID="7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e" exitCode=0 Oct 03 14:05:45 crc kubenswrapper[4959]: I1003 14:05:45.822542 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerDied","Data":"7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e"} Oct 03 14:05:45 crc kubenswrapper[4959]: I1003 14:05:45.822946 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerStarted","Data":"94e5369337af37d3e0087a4620744e070c4a9c831af5a0a359d230a348761f63"} Oct 03 14:05:45 crc kubenswrapper[4959]: I1003 14:05:45.825475 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:05:46 crc kubenswrapper[4959]: I1003 14:05:46.833526 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerStarted","Data":"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d"} Oct 03 14:05:47 crc kubenswrapper[4959]: I1003 14:05:47.844372 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerID="c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d" exitCode=0 Oct 03 14:05:47 crc kubenswrapper[4959]: I1003 14:05:47.844437 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerDied","Data":"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d"} Oct 03 14:05:48 crc kubenswrapper[4959]: I1003 14:05:48.852529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerStarted","Data":"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6"} Oct 03 14:05:48 crc kubenswrapper[4959]: I1003 14:05:48.869116 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j7cr7" podStartSLOduration=2.435203839 podStartE2EDuration="4.869095267s" podCreationTimestamp="2025-10-03 14:05:44 +0000 UTC" firstStartedPulling="2025-10-03 14:05:45.825228142 +0000 UTC m=+2115.028571559" lastFinishedPulling="2025-10-03 14:05:48.25911956 +0000 UTC m=+2117.462462987" observedRunningTime="2025-10-03 14:05:48.868102164 +0000 UTC m=+2118.071445611" watchObservedRunningTime="2025-10-03 14:05:48.869095267 +0000 UTC m=+2118.072438684" Oct 03 14:05:54 crc kubenswrapper[4959]: I1003 14:05:54.708183 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:54 crc kubenswrapper[4959]: I1003 14:05:54.708519 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:54 crc kubenswrapper[4959]: I1003 14:05:54.772339 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:54 crc kubenswrapper[4959]: I1003 14:05:54.978790 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:55 crc kubenswrapper[4959]: I1003 14:05:55.022041 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:56 crc kubenswrapper[4959]: I1003 14:05:56.950230 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j7cr7" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="registry-server" containerID="cri-o://08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6" gracePeriod=2 Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.295229 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.407788 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities\") pod \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.408058 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzps5\" (UniqueName: \"kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5\") pod \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.408104 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content\") pod \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\" (UID: \"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69\") " Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.408686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities" (OuterVolumeSpecName: "utilities") pod "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" (UID: "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.422409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5" (OuterVolumeSpecName: "kube-api-access-kzps5") pod "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" (UID: "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69"). InnerVolumeSpecName "kube-api-access-kzps5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.428481 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" (UID: "ea9cd20a-aab7-4a6c-a69c-8b2e73487e69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.509588 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.509626 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzps5\" (UniqueName: \"kubernetes.io/projected/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-kube-api-access-kzps5\") on node \"crc\" DevicePath \"\"" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.509640 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.960090 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerID="08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6" exitCode=0 Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.960140 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerDied","Data":"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6"} Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.960161 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j7cr7" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.960175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j7cr7" event={"ID":"ea9cd20a-aab7-4a6c-a69c-8b2e73487e69","Type":"ContainerDied","Data":"94e5369337af37d3e0087a4620744e070c4a9c831af5a0a359d230a348761f63"} Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.960230 4959 scope.go:117] "RemoveContainer" containerID="08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.987665 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.988157 4959 scope.go:117] "RemoveContainer" containerID="c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d" Oct 03 14:05:57 crc kubenswrapper[4959]: I1003 14:05:57.993169 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j7cr7"] Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.013582 4959 scope.go:117] "RemoveContainer" containerID="7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.031639 4959 scope.go:117] "RemoveContainer" containerID="08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6" Oct 03 14:05:58 crc kubenswrapper[4959]: E1003 14:05:58.032077 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6\": container with ID starting with 08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6 not found: ID does not exist" containerID="08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.032113 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6"} err="failed to get container status \"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6\": rpc error: code = NotFound desc = could not find container \"08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6\": container with ID starting with 08fa47daca1c56550af179f1c3e09a2535a6f745ac7549ffb7f321eedef08ca6 not found: ID does not exist" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.032136 4959 scope.go:117] "RemoveContainer" containerID="c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d" Oct 03 14:05:58 crc kubenswrapper[4959]: E1003 14:05:58.032448 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d\": container with ID starting with c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d not found: ID does not exist" containerID="c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.032512 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d"} err="failed to get container status \"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d\": rpc error: code = NotFound desc = could not find container \"c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d\": container with ID starting with c67b404d52e30c88bbecf1362cd2ace7fc6adb835e91532c9fb5bd01a536984d not found: ID does not exist" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.032526 4959 scope.go:117] "RemoveContainer" containerID="7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e" Oct 03 14:05:58 crc kubenswrapper[4959]: E1003 14:05:58.032769 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e\": container with ID starting with 7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e not found: ID does not exist" containerID="7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e" Oct 03 14:05:58 crc kubenswrapper[4959]: I1003 14:05:58.032793 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e"} err="failed to get container status \"7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e\": rpc error: code = NotFound desc = could not find container \"7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e\": container with ID starting with 7205b4bff717f40feef160450df8e3621a3ccec5bb60519eea2be1335c78695e not found: ID does not exist" Oct 03 14:05:59 crc kubenswrapper[4959]: I1003 14:05:59.695008 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" path="/var/lib/kubelet/pods/ea9cd20a-aab7-4a6c-a69c-8b2e73487e69/volumes" Oct 03 14:06:36 crc kubenswrapper[4959]: I1003 14:06:36.044994 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:06:36 crc kubenswrapper[4959]: I1003 14:06:36.045621 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:07:06 crc kubenswrapper[4959]: I1003 14:07:06.044218 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:07:06 crc kubenswrapper[4959]: I1003 14:07:06.044971 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.044621 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.045042 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.045092 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.045708 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.045776 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" gracePeriod=600 Oct 03 14:07:36 crc kubenswrapper[4959]: E1003 14:07:36.167915 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.779996 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" exitCode=0 Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.780071 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a"} Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.780367 4959 scope.go:117] "RemoveContainer" containerID="9f057211e5aa876a70bfd2cebee7da337e2ee6dc4e0a1d4f7024007443b2d7d1" Oct 03 14:07:36 crc kubenswrapper[4959]: I1003 14:07:36.780911 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:07:36 crc kubenswrapper[4959]: E1003 14:07:36.781131 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:07:47 crc kubenswrapper[4959]: I1003 14:07:47.688054 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:07:47 crc kubenswrapper[4959]: E1003 14:07:47.689170 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:07:58 crc kubenswrapper[4959]: I1003 14:07:58.685657 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:07:58 crc kubenswrapper[4959]: E1003 14:07:58.688455 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:08:10 crc kubenswrapper[4959]: I1003 14:08:10.684943 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:08:10 crc kubenswrapper[4959]: E1003 14:08:10.686640 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.864705 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:17 crc kubenswrapper[4959]: E1003 14:08:17.865404 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="registry-server" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.865416 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="registry-server" Oct 03 14:08:17 crc kubenswrapper[4959]: E1003 14:08:17.865440 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="extract-utilities" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.865446 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="extract-utilities" Oct 03 14:08:17 crc kubenswrapper[4959]: E1003 14:08:17.865456 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="extract-content" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.865462 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="extract-content" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.865598 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9cd20a-aab7-4a6c-a69c-8b2e73487e69" containerName="registry-server" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.866517 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.882188 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.930635 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjwvm\" (UniqueName: \"kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.930693 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:17 crc kubenswrapper[4959]: I1003 14:08:17.930844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.032160 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjwvm\" (UniqueName: \"kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.032226 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.032269 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.032718 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.032853 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.049552 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjwvm\" (UniqueName: \"kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm\") pod \"redhat-operators-q5blv\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.185633 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:18 crc kubenswrapper[4959]: I1003 14:08:18.680730 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:19 crc kubenswrapper[4959]: I1003 14:08:19.128173 4959 generic.go:334] "Generic (PLEG): container finished" podID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerID="429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746" exitCode=0 Oct 03 14:08:19 crc kubenswrapper[4959]: I1003 14:08:19.128259 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerDied","Data":"429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746"} Oct 03 14:08:19 crc kubenswrapper[4959]: I1003 14:08:19.128481 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerStarted","Data":"e4d208473e645f97e70eea9c560d3bb1b5b54541d87761b04fd9766a73e09a67"} Oct 03 14:08:21 crc kubenswrapper[4959]: I1003 14:08:21.144623 4959 generic.go:334] "Generic (PLEG): container finished" podID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerID="2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8" exitCode=0 Oct 03 14:08:21 crc kubenswrapper[4959]: I1003 14:08:21.144725 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerDied","Data":"2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8"} Oct 03 14:08:22 crc kubenswrapper[4959]: I1003 14:08:22.154952 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerStarted","Data":"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240"} Oct 03 14:08:22 crc kubenswrapper[4959]: I1003 14:08:22.183980 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q5blv" podStartSLOduration=2.735058695 podStartE2EDuration="5.183961666s" podCreationTimestamp="2025-10-03 14:08:17 +0000 UTC" firstStartedPulling="2025-10-03 14:08:19.129665461 +0000 UTC m=+2268.333008878" lastFinishedPulling="2025-10-03 14:08:21.578568412 +0000 UTC m=+2270.781911849" observedRunningTime="2025-10-03 14:08:22.174258769 +0000 UTC m=+2271.377602196" watchObservedRunningTime="2025-10-03 14:08:22.183961666 +0000 UTC m=+2271.387305083" Oct 03 14:08:23 crc kubenswrapper[4959]: I1003 14:08:23.685134 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:08:23 crc kubenswrapper[4959]: E1003 14:08:23.685479 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:08:28 crc kubenswrapper[4959]: I1003 14:08:28.186899 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:28 crc kubenswrapper[4959]: I1003 14:08:28.187289 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:28 crc kubenswrapper[4959]: I1003 14:08:28.229000 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:29 crc kubenswrapper[4959]: I1003 14:08:29.249026 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:29 crc kubenswrapper[4959]: I1003 14:08:29.299701 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:31 crc kubenswrapper[4959]: I1003 14:08:31.213849 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q5blv" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="registry-server" containerID="cri-o://5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240" gracePeriod=2 Oct 03 14:08:32 crc kubenswrapper[4959]: I1003 14:08:32.968264 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.143460 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjwvm\" (UniqueName: \"kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm\") pod \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.143529 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities\") pod \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.143595 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content\") pod \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\" (UID: \"91459c9d-c6a5-46cf-ae64-21dc7829fb55\") " Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.144574 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities" (OuterVolumeSpecName: "utilities") pod "91459c9d-c6a5-46cf-ae64-21dc7829fb55" (UID: "91459c9d-c6a5-46cf-ae64-21dc7829fb55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.151837 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm" (OuterVolumeSpecName: "kube-api-access-gjwvm") pod "91459c9d-c6a5-46cf-ae64-21dc7829fb55" (UID: "91459c9d-c6a5-46cf-ae64-21dc7829fb55"). InnerVolumeSpecName "kube-api-access-gjwvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.231230 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91459c9d-c6a5-46cf-ae64-21dc7829fb55" (UID: "91459c9d-c6a5-46cf-ae64-21dc7829fb55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.241382 4959 generic.go:334] "Generic (PLEG): container finished" podID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerID="5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240" exitCode=0 Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.241430 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerDied","Data":"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240"} Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.241435 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5blv" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.241456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5blv" event={"ID":"91459c9d-c6a5-46cf-ae64-21dc7829fb55","Type":"ContainerDied","Data":"e4d208473e645f97e70eea9c560d3bb1b5b54541d87761b04fd9766a73e09a67"} Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.241474 4959 scope.go:117] "RemoveContainer" containerID="5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.245072 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.245099 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjwvm\" (UniqueName: \"kubernetes.io/projected/91459c9d-c6a5-46cf-ae64-21dc7829fb55-kube-api-access-gjwvm\") on node \"crc\" DevicePath \"\"" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.245108 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91459c9d-c6a5-46cf-ae64-21dc7829fb55-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.259268 4959 scope.go:117] "RemoveContainer" containerID="2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.273881 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.279533 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q5blv"] Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.294801 4959 scope.go:117] "RemoveContainer" containerID="429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.318942 4959 scope.go:117] "RemoveContainer" containerID="5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240" Oct 03 14:08:33 crc kubenswrapper[4959]: E1003 14:08:33.319375 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240\": container with ID starting with 5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240 not found: ID does not exist" containerID="5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.319424 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240"} err="failed to get container status \"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240\": rpc error: code = NotFound desc = could not find container \"5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240\": container with ID starting with 5db0b949e3d8e21a52e46ec73ac53a6d3bba4467e4debdae34f982a167892240 not found: ID does not exist" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.319476 4959 scope.go:117] "RemoveContainer" containerID="2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8" Oct 03 14:08:33 crc kubenswrapper[4959]: E1003 14:08:33.319813 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8\": container with ID starting with 2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8 not found: ID does not exist" containerID="2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.319851 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8"} err="failed to get container status \"2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8\": rpc error: code = NotFound desc = could not find container \"2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8\": container with ID starting with 2f39b9be7a2fcade4552dbcaba106f227b75b671e85667e0a66e4a229560a3b8 not found: ID does not exist" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.319877 4959 scope.go:117] "RemoveContainer" containerID="429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746" Oct 03 14:08:33 crc kubenswrapper[4959]: E1003 14:08:33.320152 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746\": container with ID starting with 429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746 not found: ID does not exist" containerID="429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.320176 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746"} err="failed to get container status \"429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746\": rpc error: code = NotFound desc = could not find container \"429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746\": container with ID starting with 429945f3ef45baa964ddd3aed9be4734464dc2ba8aa554d02d42cbff2e5ca746 not found: ID does not exist" Oct 03 14:08:33 crc kubenswrapper[4959]: I1003 14:08:33.694773 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" path="/var/lib/kubelet/pods/91459c9d-c6a5-46cf-ae64-21dc7829fb55/volumes" Oct 03 14:08:36 crc kubenswrapper[4959]: I1003 14:08:36.685984 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:08:36 crc kubenswrapper[4959]: E1003 14:08:36.686753 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:08:51 crc kubenswrapper[4959]: I1003 14:08:51.690175 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:08:51 crc kubenswrapper[4959]: E1003 14:08:51.691075 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:09:02 crc kubenswrapper[4959]: I1003 14:09:02.685822 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:09:02 crc kubenswrapper[4959]: E1003 14:09:02.687003 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:09:13 crc kubenswrapper[4959]: I1003 14:09:13.685745 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:09:13 crc kubenswrapper[4959]: E1003 14:09:13.686647 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:09:26 crc kubenswrapper[4959]: I1003 14:09:26.686380 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:09:26 crc kubenswrapper[4959]: E1003 14:09:26.689626 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:09:39 crc kubenswrapper[4959]: I1003 14:09:39.684944 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:09:39 crc kubenswrapper[4959]: E1003 14:09:39.685677 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:09:52 crc kubenswrapper[4959]: I1003 14:09:52.684990 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:09:52 crc kubenswrapper[4959]: E1003 14:09:52.686114 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:10:06 crc kubenswrapper[4959]: I1003 14:10:06.686291 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:10:06 crc kubenswrapper[4959]: E1003 14:10:06.688143 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:10:21 crc kubenswrapper[4959]: I1003 14:10:21.690050 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:10:21 crc kubenswrapper[4959]: E1003 14:10:21.691918 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.681885 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:26 crc kubenswrapper[4959]: E1003 14:10:26.682531 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="extract-content" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.682546 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="extract-content" Oct 03 14:10:26 crc kubenswrapper[4959]: E1003 14:10:26.682560 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="registry-server" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.682566 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="registry-server" Oct 03 14:10:26 crc kubenswrapper[4959]: E1003 14:10:26.682588 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="extract-utilities" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.682596 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="extract-utilities" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.682735 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="91459c9d-c6a5-46cf-ae64-21dc7829fb55" containerName="registry-server" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.683705 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.688964 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.718349 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.718454 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.718569 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8bks\" (UniqueName: \"kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.819616 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.819715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8bks\" (UniqueName: \"kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.819771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.820185 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.820252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:26 crc kubenswrapper[4959]: I1003 14:10:26.838339 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8bks\" (UniqueName: \"kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks\") pod \"certified-operators-6vkqs\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:27 crc kubenswrapper[4959]: I1003 14:10:27.034017 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:27 crc kubenswrapper[4959]: I1003 14:10:27.503494 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:27 crc kubenswrapper[4959]: W1003 14:10:27.510600 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ef9b4de_ee7b_42a9_98e4_ca06670a6f16.slice/crio-64e1d3ae39535551f172be7039e5f4948bb45609043ae1a2d5b5f379d318b0f1 WatchSource:0}: Error finding container 64e1d3ae39535551f172be7039e5f4948bb45609043ae1a2d5b5f379d318b0f1: Status 404 returned error can't find the container with id 64e1d3ae39535551f172be7039e5f4948bb45609043ae1a2d5b5f379d318b0f1 Oct 03 14:10:28 crc kubenswrapper[4959]: I1003 14:10:28.135093 4959 generic.go:334] "Generic (PLEG): container finished" podID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerID="5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945" exitCode=0 Oct 03 14:10:28 crc kubenswrapper[4959]: I1003 14:10:28.135173 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerDied","Data":"5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945"} Oct 03 14:10:28 crc kubenswrapper[4959]: I1003 14:10:28.135482 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerStarted","Data":"64e1d3ae39535551f172be7039e5f4948bb45609043ae1a2d5b5f379d318b0f1"} Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.078617 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.082338 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.096230 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.154630 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph2l4\" (UniqueName: \"kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.154679 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.154781 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.255858 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.256307 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph2l4\" (UniqueName: \"kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.256331 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.256867 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.256930 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.283761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph2l4\" (UniqueName: \"kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4\") pod \"community-operators-9fv64\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.451592 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:29 crc kubenswrapper[4959]: W1003 14:10:29.694451 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95f8840d_c283_4c58_a8fb_e5c9c73e2ea1.slice/crio-a70e653448ae7243a97b687565d59d866f15404d5f5742cbbaa900c19939f1ce WatchSource:0}: Error finding container a70e653448ae7243a97b687565d59d866f15404d5f5742cbbaa900c19939f1ce: Status 404 returned error can't find the container with id a70e653448ae7243a97b687565d59d866f15404d5f5742cbbaa900c19939f1ce Oct 03 14:10:29 crc kubenswrapper[4959]: I1003 14:10:29.696172 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:30 crc kubenswrapper[4959]: I1003 14:10:30.157552 4959 generic.go:334] "Generic (PLEG): container finished" podID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerID="266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c" exitCode=0 Oct 03 14:10:30 crc kubenswrapper[4959]: I1003 14:10:30.157614 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerDied","Data":"266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c"} Oct 03 14:10:30 crc kubenswrapper[4959]: I1003 14:10:30.160050 4959 generic.go:334] "Generic (PLEG): container finished" podID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerID="2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93" exitCode=0 Oct 03 14:10:30 crc kubenswrapper[4959]: I1003 14:10:30.160086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerDied","Data":"2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93"} Oct 03 14:10:30 crc kubenswrapper[4959]: I1003 14:10:30.160112 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerStarted","Data":"a70e653448ae7243a97b687565d59d866f15404d5f5742cbbaa900c19939f1ce"} Oct 03 14:10:31 crc kubenswrapper[4959]: I1003 14:10:31.173381 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerStarted","Data":"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c"} Oct 03 14:10:31 crc kubenswrapper[4959]: I1003 14:10:31.197541 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6vkqs" podStartSLOduration=2.516941459 podStartE2EDuration="5.197510735s" podCreationTimestamp="2025-10-03 14:10:26 +0000 UTC" firstStartedPulling="2025-10-03 14:10:28.137144831 +0000 UTC m=+2397.340488248" lastFinishedPulling="2025-10-03 14:10:30.817714107 +0000 UTC m=+2400.021057524" observedRunningTime="2025-10-03 14:10:31.196154692 +0000 UTC m=+2400.399498119" watchObservedRunningTime="2025-10-03 14:10:31.197510735 +0000 UTC m=+2400.400854192" Oct 03 14:10:32 crc kubenswrapper[4959]: I1003 14:10:32.184835 4959 generic.go:334] "Generic (PLEG): container finished" podID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerID="77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0" exitCode=0 Oct 03 14:10:32 crc kubenswrapper[4959]: I1003 14:10:32.184965 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerDied","Data":"77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0"} Oct 03 14:10:33 crc kubenswrapper[4959]: I1003 14:10:33.194141 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerStarted","Data":"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a"} Oct 03 14:10:33 crc kubenswrapper[4959]: I1003 14:10:33.213515 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9fv64" podStartSLOduration=1.747673204 podStartE2EDuration="4.213495298s" podCreationTimestamp="2025-10-03 14:10:29 +0000 UTC" firstStartedPulling="2025-10-03 14:10:30.161901009 +0000 UTC m=+2399.365244436" lastFinishedPulling="2025-10-03 14:10:32.627723103 +0000 UTC m=+2401.831066530" observedRunningTime="2025-10-03 14:10:33.213226781 +0000 UTC m=+2402.416570208" watchObservedRunningTime="2025-10-03 14:10:33.213495298 +0000 UTC m=+2402.416838715" Oct 03 14:10:35 crc kubenswrapper[4959]: I1003 14:10:35.685762 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:10:35 crc kubenswrapper[4959]: E1003 14:10:35.686238 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:10:37 crc kubenswrapper[4959]: I1003 14:10:37.034877 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:37 crc kubenswrapper[4959]: I1003 14:10:37.035005 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:37 crc kubenswrapper[4959]: I1003 14:10:37.081808 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:37 crc kubenswrapper[4959]: I1003 14:10:37.277553 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:37 crc kubenswrapper[4959]: I1003 14:10:37.865427 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.236900 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6vkqs" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="registry-server" containerID="cri-o://f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c" gracePeriod=2 Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.452324 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.452437 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.512645 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.685864 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.820303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities\") pod \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.820379 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content\") pod \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.820482 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8bks\" (UniqueName: \"kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks\") pod \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\" (UID: \"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16\") " Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.821184 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities" (OuterVolumeSpecName: "utilities") pod "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" (UID: "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.826747 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks" (OuterVolumeSpecName: "kube-api-access-k8bks") pod "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" (UID: "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16"). InnerVolumeSpecName "kube-api-access-k8bks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.922622 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8bks\" (UniqueName: \"kubernetes.io/projected/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-kube-api-access-k8bks\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:39 crc kubenswrapper[4959]: I1003 14:10:39.922674 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.118407 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" (UID: "3ef9b4de-ee7b-42a9-98e4-ca06670a6f16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.126342 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.245320 4959 generic.go:334] "Generic (PLEG): container finished" podID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerID="f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c" exitCode=0 Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.245388 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6vkqs" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.245453 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerDied","Data":"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c"} Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.245505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6vkqs" event={"ID":"3ef9b4de-ee7b-42a9-98e4-ca06670a6f16","Type":"ContainerDied","Data":"64e1d3ae39535551f172be7039e5f4948bb45609043ae1a2d5b5f379d318b0f1"} Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.245527 4959 scope.go:117] "RemoveContainer" containerID="f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.267276 4959 scope.go:117] "RemoveContainer" containerID="266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.281271 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.285961 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6vkqs"] Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.287991 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.314899 4959 scope.go:117] "RemoveContainer" containerID="5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.331062 4959 scope.go:117] "RemoveContainer" containerID="f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c" Oct 03 14:10:40 crc kubenswrapper[4959]: E1003 14:10:40.331657 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c\": container with ID starting with f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c not found: ID does not exist" containerID="f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.331696 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c"} err="failed to get container status \"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c\": rpc error: code = NotFound desc = could not find container \"f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c\": container with ID starting with f7546cecaa5cd7526119010beb13a268935cba90b311f5e0e78be7db1760d85c not found: ID does not exist" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.331718 4959 scope.go:117] "RemoveContainer" containerID="266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c" Oct 03 14:10:40 crc kubenswrapper[4959]: E1003 14:10:40.332143 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c\": container with ID starting with 266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c not found: ID does not exist" containerID="266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.332207 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c"} err="failed to get container status \"266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c\": rpc error: code = NotFound desc = could not find container \"266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c\": container with ID starting with 266c597bc35c467d47d6a596599ad9a169c93546f3b66d2697511617c098381c not found: ID does not exist" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.332239 4959 scope.go:117] "RemoveContainer" containerID="5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945" Oct 03 14:10:40 crc kubenswrapper[4959]: E1003 14:10:40.332576 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945\": container with ID starting with 5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945 not found: ID does not exist" containerID="5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945" Oct 03 14:10:40 crc kubenswrapper[4959]: I1003 14:10:40.332612 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945"} err="failed to get container status \"5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945\": rpc error: code = NotFound desc = could not find container \"5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945\": container with ID starting with 5f26a5c54499a3cb91c2a6d65b4f271120604c36e17fc96c935e2cd5db13c945 not found: ID does not exist" Oct 03 14:10:41 crc kubenswrapper[4959]: I1003 14:10:41.701852 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" path="/var/lib/kubelet/pods/3ef9b4de-ee7b-42a9-98e4-ca06670a6f16/volumes" Oct 03 14:10:42 crc kubenswrapper[4959]: I1003 14:10:42.659794 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.274698 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9fv64" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="registry-server" containerID="cri-o://4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a" gracePeriod=2 Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.675365 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.777864 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content\") pod \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.777911 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph2l4\" (UniqueName: \"kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4\") pod \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.777955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities\") pod \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\" (UID: \"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1\") " Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.779906 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities" (OuterVolumeSpecName: "utilities") pod "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" (UID: "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.784031 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4" (OuterVolumeSpecName: "kube-api-access-ph2l4") pod "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" (UID: "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1"). InnerVolumeSpecName "kube-api-access-ph2l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.826150 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" (UID: "95f8840d-c283-4c58-a8fb-e5c9c73e2ea1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.879878 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph2l4\" (UniqueName: \"kubernetes.io/projected/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-kube-api-access-ph2l4\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.879919 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:43 crc kubenswrapper[4959]: I1003 14:10:43.879932 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.286364 4959 generic.go:334] "Generic (PLEG): container finished" podID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerID="4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a" exitCode=0 Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.286383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerDied","Data":"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a"} Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.286417 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9fv64" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.286439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9fv64" event={"ID":"95f8840d-c283-4c58-a8fb-e5c9c73e2ea1","Type":"ContainerDied","Data":"a70e653448ae7243a97b687565d59d866f15404d5f5742cbbaa900c19939f1ce"} Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.286464 4959 scope.go:117] "RemoveContainer" containerID="4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.307268 4959 scope.go:117] "RemoveContainer" containerID="77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.330698 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.339553 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9fv64"] Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.345455 4959 scope.go:117] "RemoveContainer" containerID="2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.371835 4959 scope.go:117] "RemoveContainer" containerID="4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a" Oct 03 14:10:44 crc kubenswrapper[4959]: E1003 14:10:44.372169 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a\": container with ID starting with 4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a not found: ID does not exist" containerID="4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.372218 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a"} err="failed to get container status \"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a\": rpc error: code = NotFound desc = could not find container \"4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a\": container with ID starting with 4f26bd71424aa94cf15a8b1a91682262a1f51da8e06ff9b8b3f1de32dade068a not found: ID does not exist" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.372248 4959 scope.go:117] "RemoveContainer" containerID="77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0" Oct 03 14:10:44 crc kubenswrapper[4959]: E1003 14:10:44.372868 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0\": container with ID starting with 77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0 not found: ID does not exist" containerID="77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.372912 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0"} err="failed to get container status \"77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0\": rpc error: code = NotFound desc = could not find container \"77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0\": container with ID starting with 77da0df8efbee288969fc9410d8b530931a49194250c354189afa75171ae76f0 not found: ID does not exist" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.372940 4959 scope.go:117] "RemoveContainer" containerID="2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93" Oct 03 14:10:44 crc kubenswrapper[4959]: E1003 14:10:44.373165 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93\": container with ID starting with 2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93 not found: ID does not exist" containerID="2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93" Oct 03 14:10:44 crc kubenswrapper[4959]: I1003 14:10:44.373217 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93"} err="failed to get container status \"2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93\": rpc error: code = NotFound desc = could not find container \"2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93\": container with ID starting with 2563d42824ff21cd894cb95ec3220380990a4386d92898423e1dc74292a2fb93 not found: ID does not exist" Oct 03 14:10:45 crc kubenswrapper[4959]: I1003 14:10:45.695088 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" path="/var/lib/kubelet/pods/95f8840d-c283-4c58-a8fb-e5c9c73e2ea1/volumes" Oct 03 14:10:47 crc kubenswrapper[4959]: I1003 14:10:47.686245 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:10:47 crc kubenswrapper[4959]: E1003 14:10:47.686941 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:11:01 crc kubenswrapper[4959]: I1003 14:11:01.697435 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:11:01 crc kubenswrapper[4959]: E1003 14:11:01.699278 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:11:13 crc kubenswrapper[4959]: I1003 14:11:13.685358 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:11:13 crc kubenswrapper[4959]: E1003 14:11:13.686241 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:11:28 crc kubenswrapper[4959]: I1003 14:11:28.685526 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:11:28 crc kubenswrapper[4959]: E1003 14:11:28.686511 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:11:42 crc kubenswrapper[4959]: I1003 14:11:42.685796 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:11:42 crc kubenswrapper[4959]: E1003 14:11:42.686712 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:11:57 crc kubenswrapper[4959]: I1003 14:11:57.685999 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:11:57 crc kubenswrapper[4959]: E1003 14:11:57.686689 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:12:11 crc kubenswrapper[4959]: I1003 14:12:11.692847 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:12:11 crc kubenswrapper[4959]: E1003 14:12:11.693612 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:12:22 crc kubenswrapper[4959]: I1003 14:12:22.685025 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:12:22 crc kubenswrapper[4959]: E1003 14:12:22.685759 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:12:37 crc kubenswrapper[4959]: I1003 14:12:37.685621 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:12:38 crc kubenswrapper[4959]: I1003 14:12:38.249506 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33"} Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.154641 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx"] Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155691 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155709 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155728 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155735 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155754 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155761 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155778 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155784 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155796 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155803 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: E1003 14:15:00.155813 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155819 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.155976 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef9b4de-ee7b-42a9-98e4-ca06670a6f16" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.156089 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f8840d-c283-4c58-a8fb-e5c9c73e2ea1" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.156753 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.158792 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.159713 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.171995 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx"] Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.270749 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.270808 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.270921 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.371885 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.371965 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.371992 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.372887 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.385252 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.389505 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx\") pod \"collect-profiles-29325015-7swlx\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.517961 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:00 crc kubenswrapper[4959]: I1003 14:15:00.922488 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx"] Oct 03 14:15:01 crc kubenswrapper[4959]: I1003 14:15:01.329103 4959 generic.go:334] "Generic (PLEG): container finished" podID="5ef78006-948e-4e90-ae43-798fd1f38b03" containerID="9b1a0cb0860a381daec0aa0dd4e4bc1f2c42598e3e0916acd4022f2553168334" exitCode=0 Oct 03 14:15:01 crc kubenswrapper[4959]: I1003 14:15:01.329148 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" event={"ID":"5ef78006-948e-4e90-ae43-798fd1f38b03","Type":"ContainerDied","Data":"9b1a0cb0860a381daec0aa0dd4e4bc1f2c42598e3e0916acd4022f2553168334"} Oct 03 14:15:01 crc kubenswrapper[4959]: I1003 14:15:01.329180 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" event={"ID":"5ef78006-948e-4e90-ae43-798fd1f38b03","Type":"ContainerStarted","Data":"be26485309212a476d361e387c77dfc43151b629265e1900d84786ee196d1d95"} Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.662158 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.808933 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx\") pod \"5ef78006-948e-4e90-ae43-798fd1f38b03\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.809131 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume\") pod \"5ef78006-948e-4e90-ae43-798fd1f38b03\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.809391 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume\") pod \"5ef78006-948e-4e90-ae43-798fd1f38b03\" (UID: \"5ef78006-948e-4e90-ae43-798fd1f38b03\") " Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.810103 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume" (OuterVolumeSpecName: "config-volume") pod "5ef78006-948e-4e90-ae43-798fd1f38b03" (UID: "5ef78006-948e-4e90-ae43-798fd1f38b03"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.817496 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5ef78006-948e-4e90-ae43-798fd1f38b03" (UID: "5ef78006-948e-4e90-ae43-798fd1f38b03"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.817571 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx" (OuterVolumeSpecName: "kube-api-access-h2rrx") pod "5ef78006-948e-4e90-ae43-798fd1f38b03" (UID: "5ef78006-948e-4e90-ae43-798fd1f38b03"). InnerVolumeSpecName "kube-api-access-h2rrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.910987 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ef78006-948e-4e90-ae43-798fd1f38b03-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.911034 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ef78006-948e-4e90-ae43-798fd1f38b03-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:02 crc kubenswrapper[4959]: I1003 14:15:02.911053 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2rrx\" (UniqueName: \"kubernetes.io/projected/5ef78006-948e-4e90-ae43-798fd1f38b03-kube-api-access-h2rrx\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:03 crc kubenswrapper[4959]: I1003 14:15:03.345694 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" event={"ID":"5ef78006-948e-4e90-ae43-798fd1f38b03","Type":"ContainerDied","Data":"be26485309212a476d361e387c77dfc43151b629265e1900d84786ee196d1d95"} Oct 03 14:15:03 crc kubenswrapper[4959]: I1003 14:15:03.345745 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be26485309212a476d361e387c77dfc43151b629265e1900d84786ee196d1d95" Oct 03 14:15:03 crc kubenswrapper[4959]: I1003 14:15:03.346009 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx" Oct 03 14:15:03 crc kubenswrapper[4959]: I1003 14:15:03.744435 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd"] Oct 03 14:15:03 crc kubenswrapper[4959]: I1003 14:15:03.750853 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-2gdjd"] Oct 03 14:15:05 crc kubenswrapper[4959]: I1003 14:15:05.695769 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6c97e09-88e1-4f2b-a476-19796c3d83f2" path="/var/lib/kubelet/pods/c6c97e09-88e1-4f2b-a476-19796c3d83f2/volumes" Oct 03 14:15:06 crc kubenswrapper[4959]: I1003 14:15:06.045004 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:15:06 crc kubenswrapper[4959]: I1003 14:15:06.045071 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:15:36 crc kubenswrapper[4959]: I1003 14:15:36.044562 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:15:36 crc kubenswrapper[4959]: I1003 14:15:36.045103 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:16:03 crc kubenswrapper[4959]: I1003 14:16:03.909648 4959 scope.go:117] "RemoveContainer" containerID="929bde24737df2c8c759f0b00eb1fbc144bfc33e1594131a41305ab4ac1908fd" Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.045005 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.046457 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.046547 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.047344 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.047478 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33" gracePeriod=600 Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.851862 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33" exitCode=0 Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.852214 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33"} Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.852243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24"} Oct 03 14:16:06 crc kubenswrapper[4959]: I1003 14:16:06.852258 4959 scope.go:117] "RemoveContainer" containerID="0a1ea36451957cc862469c175bbd7d8a6eedf61fe10b997a5bf37071a9e4542a" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.194333 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:01 crc kubenswrapper[4959]: E1003 14:17:01.195327 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef78006-948e-4e90-ae43-798fd1f38b03" containerName="collect-profiles" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.195344 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef78006-948e-4e90-ae43-798fd1f38b03" containerName="collect-profiles" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.195533 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef78006-948e-4e90-ae43-798fd1f38b03" containerName="collect-profiles" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.196857 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.217950 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.354058 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m92n4\" (UniqueName: \"kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.354125 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.354181 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.455234 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.455315 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m92n4\" (UniqueName: \"kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.455365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.456147 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.456257 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.475673 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m92n4\" (UniqueName: \"kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4\") pod \"redhat-marketplace-tpkp9\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.519728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:01 crc kubenswrapper[4959]: I1003 14:17:01.946979 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:02 crc kubenswrapper[4959]: I1003 14:17:02.292717 4959 generic.go:334] "Generic (PLEG): container finished" podID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerID="dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a" exitCode=0 Oct 03 14:17:02 crc kubenswrapper[4959]: I1003 14:17:02.292774 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerDied","Data":"dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a"} Oct 03 14:17:02 crc kubenswrapper[4959]: I1003 14:17:02.292809 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerStarted","Data":"b6776ea5258dcc24b7fefca21683dda50fe3b085c9bc1f85fc8c07fced95740a"} Oct 03 14:17:02 crc kubenswrapper[4959]: I1003 14:17:02.294553 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:17:03 crc kubenswrapper[4959]: I1003 14:17:03.303558 4959 generic.go:334] "Generic (PLEG): container finished" podID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerID="b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a" exitCode=0 Oct 03 14:17:03 crc kubenswrapper[4959]: I1003 14:17:03.303840 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerDied","Data":"b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a"} Oct 03 14:17:04 crc kubenswrapper[4959]: I1003 14:17:04.312671 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerStarted","Data":"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a"} Oct 03 14:17:04 crc kubenswrapper[4959]: I1003 14:17:04.344466 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tpkp9" podStartSLOduration=1.892011789 podStartE2EDuration="3.344444231s" podCreationTimestamp="2025-10-03 14:17:01 +0000 UTC" firstStartedPulling="2025-10-03 14:17:02.294306364 +0000 UTC m=+2791.497649781" lastFinishedPulling="2025-10-03 14:17:03.746738806 +0000 UTC m=+2792.950082223" observedRunningTime="2025-10-03 14:17:04.330979762 +0000 UTC m=+2793.534323189" watchObservedRunningTime="2025-10-03 14:17:04.344444231 +0000 UTC m=+2793.547787668" Oct 03 14:17:11 crc kubenswrapper[4959]: I1003 14:17:11.520115 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:11 crc kubenswrapper[4959]: I1003 14:17:11.520861 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:11 crc kubenswrapper[4959]: I1003 14:17:11.575135 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:12 crc kubenswrapper[4959]: I1003 14:17:12.451488 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:12 crc kubenswrapper[4959]: I1003 14:17:12.502815 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.387654 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tpkp9" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="registry-server" containerID="cri-o://45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a" gracePeriod=2 Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.788470 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.845935 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities\") pod \"731de539-8c98-40d2-8f3f-d2358d39c3c4\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.845971 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content\") pod \"731de539-8c98-40d2-8f3f-d2358d39c3c4\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.846024 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m92n4\" (UniqueName: \"kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4\") pod \"731de539-8c98-40d2-8f3f-d2358d39c3c4\" (UID: \"731de539-8c98-40d2-8f3f-d2358d39c3c4\") " Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.847930 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities" (OuterVolumeSpecName: "utilities") pod "731de539-8c98-40d2-8f3f-d2358d39c3c4" (UID: "731de539-8c98-40d2-8f3f-d2358d39c3c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.851026 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4" (OuterVolumeSpecName: "kube-api-access-m92n4") pod "731de539-8c98-40d2-8f3f-d2358d39c3c4" (UID: "731de539-8c98-40d2-8f3f-d2358d39c3c4"). InnerVolumeSpecName "kube-api-access-m92n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.860097 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "731de539-8c98-40d2-8f3f-d2358d39c3c4" (UID: "731de539-8c98-40d2-8f3f-d2358d39c3c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.948089 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.948122 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/731de539-8c98-40d2-8f3f-d2358d39c3c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:14 crc kubenswrapper[4959]: I1003 14:17:14.948132 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m92n4\" (UniqueName: \"kubernetes.io/projected/731de539-8c98-40d2-8f3f-d2358d39c3c4-kube-api-access-m92n4\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.398305 4959 generic.go:334] "Generic (PLEG): container finished" podID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerID="45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a" exitCode=0 Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.398374 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerDied","Data":"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a"} Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.398397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpkp9" event={"ID":"731de539-8c98-40d2-8f3f-d2358d39c3c4","Type":"ContainerDied","Data":"b6776ea5258dcc24b7fefca21683dda50fe3b085c9bc1f85fc8c07fced95740a"} Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.398413 4959 scope.go:117] "RemoveContainer" containerID="45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.398419 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpkp9" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.415333 4959 scope.go:117] "RemoveContainer" containerID="b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.438995 4959 scope.go:117] "RemoveContainer" containerID="dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.465912 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.470730 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpkp9"] Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.473177 4959 scope.go:117] "RemoveContainer" containerID="45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a" Oct 03 14:17:15 crc kubenswrapper[4959]: E1003 14:17:15.473854 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a\": container with ID starting with 45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a not found: ID does not exist" containerID="45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.473885 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a"} err="failed to get container status \"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a\": rpc error: code = NotFound desc = could not find container \"45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a\": container with ID starting with 45f2b4c1b0d8cfddb52f0ab26088bad7345526f12113d9ce7935129325fb9c9a not found: ID does not exist" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.473905 4959 scope.go:117] "RemoveContainer" containerID="b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a" Oct 03 14:17:15 crc kubenswrapper[4959]: E1003 14:17:15.474307 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a\": container with ID starting with b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a not found: ID does not exist" containerID="b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.474333 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a"} err="failed to get container status \"b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a\": rpc error: code = NotFound desc = could not find container \"b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a\": container with ID starting with b88ea415e1acfc4bcf31cea4dc5e54f5ea70493cc30e1ea7fe448c203a634b9a not found: ID does not exist" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.474347 4959 scope.go:117] "RemoveContainer" containerID="dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a" Oct 03 14:17:15 crc kubenswrapper[4959]: E1003 14:17:15.474661 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a\": container with ID starting with dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a not found: ID does not exist" containerID="dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.474686 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a"} err="failed to get container status \"dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a\": rpc error: code = NotFound desc = could not find container \"dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a\": container with ID starting with dbe21203976dd2f43f4f4c67065ae5609691c059f21bb3e1ad8694f64514664a not found: ID does not exist" Oct 03 14:17:15 crc kubenswrapper[4959]: I1003 14:17:15.693905 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" path="/var/lib/kubelet/pods/731de539-8c98-40d2-8f3f-d2358d39c3c4/volumes" Oct 03 14:18:06 crc kubenswrapper[4959]: I1003 14:18:06.044867 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:18:06 crc kubenswrapper[4959]: I1003 14:18:06.046490 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.813036 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:21 crc kubenswrapper[4959]: E1003 14:18:21.814364 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="extract-utilities" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.814384 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="extract-utilities" Oct 03 14:18:21 crc kubenswrapper[4959]: E1003 14:18:21.814414 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="registry-server" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.814421 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="registry-server" Oct 03 14:18:21 crc kubenswrapper[4959]: E1003 14:18:21.814432 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="extract-content" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.814438 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="extract-content" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.814627 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="731de539-8c98-40d2-8f3f-d2358d39c3c4" containerName="registry-server" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.816033 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.826970 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.946678 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.946751 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:21 crc kubenswrapper[4959]: I1003 14:18:21.946782 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c852\" (UniqueName: \"kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.047839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.047914 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c852\" (UniqueName: \"kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.048025 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.049805 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.049833 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.071360 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c852\" (UniqueName: \"kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852\") pod \"redhat-operators-4xgdk\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.137932 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.633925 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.920061 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerStarted","Data":"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd"} Oct 03 14:18:22 crc kubenswrapper[4959]: I1003 14:18:22.920132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerStarted","Data":"6da60ee9f62635d91f27ebc0d4e47e5cd800d223fd0a8568313cd21ab4e5af12"} Oct 03 14:18:23 crc kubenswrapper[4959]: I1003 14:18:23.930220 4959 generic.go:334] "Generic (PLEG): container finished" podID="345abfde-b272-4eee-8079-03526ecf48da" containerID="4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd" exitCode=0 Oct 03 14:18:23 crc kubenswrapper[4959]: I1003 14:18:23.930285 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerDied","Data":"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd"} Oct 03 14:18:25 crc kubenswrapper[4959]: I1003 14:18:25.945041 4959 generic.go:334] "Generic (PLEG): container finished" podID="345abfde-b272-4eee-8079-03526ecf48da" containerID="ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c" exitCode=0 Oct 03 14:18:25 crc kubenswrapper[4959]: I1003 14:18:25.945137 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerDied","Data":"ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c"} Oct 03 14:18:26 crc kubenswrapper[4959]: I1003 14:18:26.954711 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerStarted","Data":"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7"} Oct 03 14:18:26 crc kubenswrapper[4959]: I1003 14:18:26.988138 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4xgdk" podStartSLOduration=3.4748542479999998 podStartE2EDuration="5.98811143s" podCreationTimestamp="2025-10-03 14:18:21 +0000 UTC" firstStartedPulling="2025-10-03 14:18:23.932165522 +0000 UTC m=+2873.135508979" lastFinishedPulling="2025-10-03 14:18:26.445422704 +0000 UTC m=+2875.648766161" observedRunningTime="2025-10-03 14:18:26.981129719 +0000 UTC m=+2876.184473156" watchObservedRunningTime="2025-10-03 14:18:26.98811143 +0000 UTC m=+2876.191454887" Oct 03 14:18:32 crc kubenswrapper[4959]: I1003 14:18:32.138464 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:32 crc kubenswrapper[4959]: I1003 14:18:32.138911 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:32 crc kubenswrapper[4959]: I1003 14:18:32.214856 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:33 crc kubenswrapper[4959]: I1003 14:18:33.048022 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:33 crc kubenswrapper[4959]: I1003 14:18:33.093626 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.016265 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4xgdk" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="registry-server" containerID="cri-o://4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7" gracePeriod=2 Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.441376 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.534930 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities\") pod \"345abfde-b272-4eee-8079-03526ecf48da\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.535053 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c852\" (UniqueName: \"kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852\") pod \"345abfde-b272-4eee-8079-03526ecf48da\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.535136 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content\") pod \"345abfde-b272-4eee-8079-03526ecf48da\" (UID: \"345abfde-b272-4eee-8079-03526ecf48da\") " Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.536042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities" (OuterVolumeSpecName: "utilities") pod "345abfde-b272-4eee-8079-03526ecf48da" (UID: "345abfde-b272-4eee-8079-03526ecf48da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.542756 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852" (OuterVolumeSpecName: "kube-api-access-7c852") pod "345abfde-b272-4eee-8079-03526ecf48da" (UID: "345abfde-b272-4eee-8079-03526ecf48da"). InnerVolumeSpecName "kube-api-access-7c852". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.635222 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "345abfde-b272-4eee-8079-03526ecf48da" (UID: "345abfde-b272-4eee-8079-03526ecf48da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.636717 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c852\" (UniqueName: \"kubernetes.io/projected/345abfde-b272-4eee-8079-03526ecf48da-kube-api-access-7c852\") on node \"crc\" DevicePath \"\"" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.636740 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:18:35 crc kubenswrapper[4959]: I1003 14:18:35.636750 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/345abfde-b272-4eee-8079-03526ecf48da-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.025015 4959 generic.go:334] "Generic (PLEG): container finished" podID="345abfde-b272-4eee-8079-03526ecf48da" containerID="4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7" exitCode=0 Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.025062 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerDied","Data":"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7"} Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.025089 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xgdk" event={"ID":"345abfde-b272-4eee-8079-03526ecf48da","Type":"ContainerDied","Data":"6da60ee9f62635d91f27ebc0d4e47e5cd800d223fd0a8568313cd21ab4e5af12"} Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.025094 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xgdk" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.025108 4959 scope.go:117] "RemoveContainer" containerID="4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.044910 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.044973 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.050288 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.050781 4959 scope.go:117] "RemoveContainer" containerID="ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.057105 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4xgdk"] Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.074011 4959 scope.go:117] "RemoveContainer" containerID="4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.097091 4959 scope.go:117] "RemoveContainer" containerID="4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7" Oct 03 14:18:36 crc kubenswrapper[4959]: E1003 14:18:36.097615 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7\": container with ID starting with 4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7 not found: ID does not exist" containerID="4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.097650 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7"} err="failed to get container status \"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7\": rpc error: code = NotFound desc = could not find container \"4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7\": container with ID starting with 4bc404d8a359620eb4f4dbddbbe92495a2c1c9f95c8f48870e201e814a223ef7 not found: ID does not exist" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.097695 4959 scope.go:117] "RemoveContainer" containerID="ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c" Oct 03 14:18:36 crc kubenswrapper[4959]: E1003 14:18:36.098183 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c\": container with ID starting with ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c not found: ID does not exist" containerID="ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.098293 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c"} err="failed to get container status \"ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c\": rpc error: code = NotFound desc = could not find container \"ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c\": container with ID starting with ddcf0845df3c1def98981f9af7575969b07b78450732d5ab107eb0cbc328b53c not found: ID does not exist" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.098330 4959 scope.go:117] "RemoveContainer" containerID="4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd" Oct 03 14:18:36 crc kubenswrapper[4959]: E1003 14:18:36.098618 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd\": container with ID starting with 4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd not found: ID does not exist" containerID="4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd" Oct 03 14:18:36 crc kubenswrapper[4959]: I1003 14:18:36.098669 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd"} err="failed to get container status \"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd\": rpc error: code = NotFound desc = could not find container \"4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd\": container with ID starting with 4ea4ecb74b1831f0c2fa4f51518048901534db3804435feb4878012d63b626bd not found: ID does not exist" Oct 03 14:18:37 crc kubenswrapper[4959]: I1003 14:18:37.695369 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345abfde-b272-4eee-8079-03526ecf48da" path="/var/lib/kubelet/pods/345abfde-b272-4eee-8079-03526ecf48da/volumes" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.044680 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.045303 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.045363 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.046053 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.046121 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" gracePeriod=600 Oct 03 14:19:06 crc kubenswrapper[4959]: E1003 14:19:06.170074 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.262366 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" exitCode=0 Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.262456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24"} Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.262589 4959 scope.go:117] "RemoveContainer" containerID="429095b73948a31741cdf36cf4275b0011dd927b4ba7b5506892070fa5014e33" Oct 03 14:19:06 crc kubenswrapper[4959]: I1003 14:19:06.263305 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:19:06 crc kubenswrapper[4959]: E1003 14:19:06.263667 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:19:19 crc kubenswrapper[4959]: I1003 14:19:19.686211 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:19:19 crc kubenswrapper[4959]: E1003 14:19:19.687048 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:19:34 crc kubenswrapper[4959]: I1003 14:19:34.685623 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:19:34 crc kubenswrapper[4959]: E1003 14:19:34.686317 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:19:46 crc kubenswrapper[4959]: I1003 14:19:46.685997 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:19:46 crc kubenswrapper[4959]: E1003 14:19:46.687308 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:19:58 crc kubenswrapper[4959]: I1003 14:19:58.685571 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:19:58 crc kubenswrapper[4959]: E1003 14:19:58.686579 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:20:12 crc kubenswrapper[4959]: I1003 14:20:12.685696 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:20:12 crc kubenswrapper[4959]: E1003 14:20:12.686356 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:20:26 crc kubenswrapper[4959]: I1003 14:20:26.685817 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:20:26 crc kubenswrapper[4959]: E1003 14:20:26.686478 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:20:40 crc kubenswrapper[4959]: I1003 14:20:40.685436 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:20:40 crc kubenswrapper[4959]: E1003 14:20:40.686375 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:20:52 crc kubenswrapper[4959]: I1003 14:20:52.686454 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:20:52 crc kubenswrapper[4959]: E1003 14:20:52.687474 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:06 crc kubenswrapper[4959]: I1003 14:21:06.685397 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:21:06 crc kubenswrapper[4959]: E1003 14:21:06.686182 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:20 crc kubenswrapper[4959]: I1003 14:21:20.685638 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:21:20 crc kubenswrapper[4959]: E1003 14:21:20.686929 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:33 crc kubenswrapper[4959]: I1003 14:21:33.686017 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:21:33 crc kubenswrapper[4959]: E1003 14:21:33.686911 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:45 crc kubenswrapper[4959]: I1003 14:21:45.685886 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:21:45 crc kubenswrapper[4959]: E1003 14:21:45.686958 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.475035 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:21:52 crc kubenswrapper[4959]: E1003 14:21:52.476216 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="extract-content" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.476241 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="extract-content" Oct 03 14:21:52 crc kubenswrapper[4959]: E1003 14:21:52.476279 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="registry-server" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.476291 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="registry-server" Oct 03 14:21:52 crc kubenswrapper[4959]: E1003 14:21:52.476308 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="extract-utilities" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.476320 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="extract-utilities" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.476705 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="345abfde-b272-4eee-8079-03526ecf48da" containerName="registry-server" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.478377 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.484088 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.663592 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.663669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kh8h\" (UniqueName: \"kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.663739 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.765074 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kh8h\" (UniqueName: \"kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.765509 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.765626 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.766079 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.766649 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.787882 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kh8h\" (UniqueName: \"kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h\") pod \"certified-operators-825mx\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:52 crc kubenswrapper[4959]: I1003 14:21:52.800522 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:21:53 crc kubenswrapper[4959]: I1003 14:21:53.278014 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:21:53 crc kubenswrapper[4959]: I1003 14:21:53.640810 4959 generic.go:334] "Generic (PLEG): container finished" podID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerID="02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320" exitCode=0 Oct 03 14:21:53 crc kubenswrapper[4959]: I1003 14:21:53.640845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerDied","Data":"02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320"} Oct 03 14:21:53 crc kubenswrapper[4959]: I1003 14:21:53.640870 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerStarted","Data":"d07a0b4abb76a4e30d270a486eecde466706ab1a8c5e7f19375b8d2031747243"} Oct 03 14:21:54 crc kubenswrapper[4959]: I1003 14:21:54.650588 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerStarted","Data":"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60"} Oct 03 14:21:55 crc kubenswrapper[4959]: I1003 14:21:55.664661 4959 generic.go:334] "Generic (PLEG): container finished" podID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerID="dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60" exitCode=0 Oct 03 14:21:55 crc kubenswrapper[4959]: I1003 14:21:55.664788 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerDied","Data":"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60"} Oct 03 14:21:56 crc kubenswrapper[4959]: I1003 14:21:56.673924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerStarted","Data":"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d"} Oct 03 14:21:56 crc kubenswrapper[4959]: I1003 14:21:56.690784 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-825mx" podStartSLOduration=2.226891042 podStartE2EDuration="4.690768495s" podCreationTimestamp="2025-10-03 14:21:52 +0000 UTC" firstStartedPulling="2025-10-03 14:21:53.642691428 +0000 UTC m=+3082.846034895" lastFinishedPulling="2025-10-03 14:21:56.106568921 +0000 UTC m=+3085.309912348" observedRunningTime="2025-10-03 14:21:56.689225467 +0000 UTC m=+3085.892568904" watchObservedRunningTime="2025-10-03 14:21:56.690768495 +0000 UTC m=+3085.894111912" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.657959 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.661687 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.669217 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.856447 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.856955 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd644\" (UniqueName: \"kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.857009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.958710 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd644\" (UniqueName: \"kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.958811 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.958851 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.959534 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.959560 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.980405 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd644\" (UniqueName: \"kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644\") pod \"community-operators-872vf\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:58 crc kubenswrapper[4959]: I1003 14:21:58.991843 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:21:59 crc kubenswrapper[4959]: I1003 14:21:59.265611 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:21:59 crc kubenswrapper[4959]: I1003 14:21:59.686243 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:21:59 crc kubenswrapper[4959]: E1003 14:21:59.686770 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:21:59 crc kubenswrapper[4959]: I1003 14:21:59.714592 4959 generic.go:334] "Generic (PLEG): container finished" podID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerID="d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb" exitCode=0 Oct 03 14:21:59 crc kubenswrapper[4959]: I1003 14:21:59.714644 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerDied","Data":"d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb"} Oct 03 14:21:59 crc kubenswrapper[4959]: I1003 14:21:59.714673 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerStarted","Data":"1eb8203af3b1d6d2bf96401335257168dea6dd6b49346e30d47832a63abba56c"} Oct 03 14:22:00 crc kubenswrapper[4959]: I1003 14:22:00.724409 4959 generic.go:334] "Generic (PLEG): container finished" podID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerID="7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db" exitCode=0 Oct 03 14:22:00 crc kubenswrapper[4959]: I1003 14:22:00.724505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerDied","Data":"7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db"} Oct 03 14:22:01 crc kubenswrapper[4959]: I1003 14:22:01.733964 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerStarted","Data":"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c"} Oct 03 14:22:01 crc kubenswrapper[4959]: I1003 14:22:01.757831 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-872vf" podStartSLOduration=2.2783051690000002 podStartE2EDuration="3.757811376s" podCreationTimestamp="2025-10-03 14:21:58 +0000 UTC" firstStartedPulling="2025-10-03 14:21:59.718294278 +0000 UTC m=+3088.921637705" lastFinishedPulling="2025-10-03 14:22:01.197800495 +0000 UTC m=+3090.401143912" observedRunningTime="2025-10-03 14:22:01.755137381 +0000 UTC m=+3090.958480798" watchObservedRunningTime="2025-10-03 14:22:01.757811376 +0000 UTC m=+3090.961154793" Oct 03 14:22:02 crc kubenswrapper[4959]: I1003 14:22:02.800982 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:02 crc kubenswrapper[4959]: I1003 14:22:02.801100 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:02 crc kubenswrapper[4959]: I1003 14:22:02.865405 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:03 crc kubenswrapper[4959]: I1003 14:22:03.801673 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:04 crc kubenswrapper[4959]: I1003 14:22:04.053171 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:22:05 crc kubenswrapper[4959]: I1003 14:22:05.762432 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-825mx" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="registry-server" containerID="cri-o://143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d" gracePeriod=2 Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.235531 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.279341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kh8h\" (UniqueName: \"kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h\") pod \"57959a41-ef30-43a2-a47d-7bc907a7be36\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.279853 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content\") pod \"57959a41-ef30-43a2-a47d-7bc907a7be36\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.280224 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities\") pod \"57959a41-ef30-43a2-a47d-7bc907a7be36\" (UID: \"57959a41-ef30-43a2-a47d-7bc907a7be36\") " Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.281009 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities" (OuterVolumeSpecName: "utilities") pod "57959a41-ef30-43a2-a47d-7bc907a7be36" (UID: "57959a41-ef30-43a2-a47d-7bc907a7be36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.288597 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h" (OuterVolumeSpecName: "kube-api-access-5kh8h") pod "57959a41-ef30-43a2-a47d-7bc907a7be36" (UID: "57959a41-ef30-43a2-a47d-7bc907a7be36"). InnerVolumeSpecName "kube-api-access-5kh8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.335895 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57959a41-ef30-43a2-a47d-7bc907a7be36" (UID: "57959a41-ef30-43a2-a47d-7bc907a7be36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.382144 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.382264 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kh8h\" (UniqueName: \"kubernetes.io/projected/57959a41-ef30-43a2-a47d-7bc907a7be36-kube-api-access-5kh8h\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.382286 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57959a41-ef30-43a2-a47d-7bc907a7be36-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.773581 4959 generic.go:334] "Generic (PLEG): container finished" podID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerID="143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d" exitCode=0 Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.773621 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerDied","Data":"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d"} Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.773645 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-825mx" event={"ID":"57959a41-ef30-43a2-a47d-7bc907a7be36","Type":"ContainerDied","Data":"d07a0b4abb76a4e30d270a486eecde466706ab1a8c5e7f19375b8d2031747243"} Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.773661 4959 scope.go:117] "RemoveContainer" containerID="143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.774828 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-825mx" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.796307 4959 scope.go:117] "RemoveContainer" containerID="dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.821383 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.825746 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-825mx"] Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.832516 4959 scope.go:117] "RemoveContainer" containerID="02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.860507 4959 scope.go:117] "RemoveContainer" containerID="143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d" Oct 03 14:22:06 crc kubenswrapper[4959]: E1003 14:22:06.861091 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d\": container with ID starting with 143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d not found: ID does not exist" containerID="143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.861164 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d"} err="failed to get container status \"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d\": rpc error: code = NotFound desc = could not find container \"143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d\": container with ID starting with 143783ee18e08cf001834ce2b74ffd11f9d8ae3b9598814df08f347a58405b8d not found: ID does not exist" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.861265 4959 scope.go:117] "RemoveContainer" containerID="dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60" Oct 03 14:22:06 crc kubenswrapper[4959]: E1003 14:22:06.861675 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60\": container with ID starting with dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60 not found: ID does not exist" containerID="dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.861712 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60"} err="failed to get container status \"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60\": rpc error: code = NotFound desc = could not find container \"dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60\": container with ID starting with dcfb4fd5b46aa4f17b358915b49f7da9139239b129dfc97299596c13ef2cff60 not found: ID does not exist" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.861739 4959 scope.go:117] "RemoveContainer" containerID="02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320" Oct 03 14:22:06 crc kubenswrapper[4959]: E1003 14:22:06.862027 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320\": container with ID starting with 02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320 not found: ID does not exist" containerID="02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320" Oct 03 14:22:06 crc kubenswrapper[4959]: I1003 14:22:06.862076 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320"} err="failed to get container status \"02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320\": rpc error: code = NotFound desc = could not find container \"02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320\": container with ID starting with 02514d68ebdc914326d15f9a2f0a62f9b6a51b9a5442433a79f77eabf2f33320 not found: ID does not exist" Oct 03 14:22:07 crc kubenswrapper[4959]: I1003 14:22:07.698271 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" path="/var/lib/kubelet/pods/57959a41-ef30-43a2-a47d-7bc907a7be36/volumes" Oct 03 14:22:08 crc kubenswrapper[4959]: I1003 14:22:08.992776 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:08 crc kubenswrapper[4959]: I1003 14:22:08.992841 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:09 crc kubenswrapper[4959]: I1003 14:22:09.045287 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:09 crc kubenswrapper[4959]: I1003 14:22:09.866287 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:10 crc kubenswrapper[4959]: I1003 14:22:10.641868 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:22:11 crc kubenswrapper[4959]: I1003 14:22:11.837489 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-872vf" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="registry-server" containerID="cri-o://865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c" gracePeriod=2 Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.251619 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.371354 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content\") pod \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.371467 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd644\" (UniqueName: \"kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644\") pod \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.371494 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities\") pod \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\" (UID: \"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5\") " Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.372500 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities" (OuterVolumeSpecName: "utilities") pod "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" (UID: "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.380443 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644" (OuterVolumeSpecName: "kube-api-access-vd644") pod "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" (UID: "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5"). InnerVolumeSpecName "kube-api-access-vd644". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.427386 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" (UID: "1c865c7a-fd27-4bd0-ba15-d6540c6f7db5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.472940 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd644\" (UniqueName: \"kubernetes.io/projected/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-kube-api-access-vd644\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.472991 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.473011 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.850170 4959 generic.go:334] "Generic (PLEG): container finished" podID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerID="865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c" exitCode=0 Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.850240 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerDied","Data":"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c"} Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.850283 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-872vf" event={"ID":"1c865c7a-fd27-4bd0-ba15-d6540c6f7db5","Type":"ContainerDied","Data":"1eb8203af3b1d6d2bf96401335257168dea6dd6b49346e30d47832a63abba56c"} Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.850305 4959 scope.go:117] "RemoveContainer" containerID="865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.850312 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-872vf" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.879279 4959 scope.go:117] "RemoveContainer" containerID="7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.902176 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.908731 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-872vf"] Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.927123 4959 scope.go:117] "RemoveContainer" containerID="d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.943140 4959 scope.go:117] "RemoveContainer" containerID="865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c" Oct 03 14:22:12 crc kubenswrapper[4959]: E1003 14:22:12.943705 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c\": container with ID starting with 865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c not found: ID does not exist" containerID="865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.943755 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c"} err="failed to get container status \"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c\": rpc error: code = NotFound desc = could not find container \"865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c\": container with ID starting with 865cd095707ff36ce1641ee5869ea9db5e0476a7162f44f1d41aba84fef9304c not found: ID does not exist" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.943788 4959 scope.go:117] "RemoveContainer" containerID="7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db" Oct 03 14:22:12 crc kubenswrapper[4959]: E1003 14:22:12.944417 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db\": container with ID starting with 7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db not found: ID does not exist" containerID="7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.944503 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db"} err="failed to get container status \"7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db\": rpc error: code = NotFound desc = could not find container \"7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db\": container with ID starting with 7689dc2a6fe46e5011d837de0ee292356219120f758fb00a42283527953158db not found: ID does not exist" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.944562 4959 scope.go:117] "RemoveContainer" containerID="d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb" Oct 03 14:22:12 crc kubenswrapper[4959]: E1003 14:22:12.945043 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb\": container with ID starting with d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb not found: ID does not exist" containerID="d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb" Oct 03 14:22:12 crc kubenswrapper[4959]: I1003 14:22:12.945098 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb"} err="failed to get container status \"d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb\": rpc error: code = NotFound desc = could not find container \"d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb\": container with ID starting with d1ce10801f48f463de30b8da6031b64492be862f9e773018ffa7b1edede927cb not found: ID does not exist" Oct 03 14:22:13 crc kubenswrapper[4959]: I1003 14:22:13.706456 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" path="/var/lib/kubelet/pods/1c865c7a-fd27-4bd0-ba15-d6540c6f7db5/volumes" Oct 03 14:22:14 crc kubenswrapper[4959]: I1003 14:22:14.686000 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:22:14 crc kubenswrapper[4959]: E1003 14:22:14.686725 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:22:27 crc kubenswrapper[4959]: I1003 14:22:27.685517 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:22:27 crc kubenswrapper[4959]: E1003 14:22:27.686241 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:22:38 crc kubenswrapper[4959]: I1003 14:22:38.685925 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:22:38 crc kubenswrapper[4959]: E1003 14:22:38.686813 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:22:50 crc kubenswrapper[4959]: I1003 14:22:50.685540 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:22:50 crc kubenswrapper[4959]: E1003 14:22:50.686764 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:23:04 crc kubenswrapper[4959]: I1003 14:23:04.685424 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:23:04 crc kubenswrapper[4959]: E1003 14:23:04.686415 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:23:15 crc kubenswrapper[4959]: I1003 14:23:15.685275 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:23:15 crc kubenswrapper[4959]: E1003 14:23:15.685957 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:23:29 crc kubenswrapper[4959]: I1003 14:23:29.686236 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:23:29 crc kubenswrapper[4959]: E1003 14:23:29.687040 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:23:41 crc kubenswrapper[4959]: I1003 14:23:41.690229 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:23:41 crc kubenswrapper[4959]: E1003 14:23:41.691733 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:23:52 crc kubenswrapper[4959]: I1003 14:23:52.685463 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:23:52 crc kubenswrapper[4959]: E1003 14:23:52.686411 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:24:07 crc kubenswrapper[4959]: I1003 14:24:07.685917 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:24:08 crc kubenswrapper[4959]: I1003 14:24:08.741905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834"} Oct 03 14:26:36 crc kubenswrapper[4959]: I1003 14:26:36.045089 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:26:36 crc kubenswrapper[4959]: I1003 14:26:36.046032 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:27:06 crc kubenswrapper[4959]: I1003 14:27:06.045295 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:27:06 crc kubenswrapper[4959]: I1003 14:27:06.046413 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.262781 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274653 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="extract-content" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274672 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="extract-content" Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274683 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274689 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274711 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="extract-utilities" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274720 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="extract-utilities" Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274734 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274744 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274796 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="extract-content" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274805 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="extract-content" Oct 03 14:27:23 crc kubenswrapper[4959]: E1003 14:27:23.274822 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="extract-utilities" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274830 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="extract-utilities" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.274995 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c865c7a-fd27-4bd0-ba15-d6540c6f7db5" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.275009 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="57959a41-ef30-43a2-a47d-7bc907a7be36" containerName="registry-server" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.276163 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.276281 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.395883 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nsfh\" (UniqueName: \"kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.396291 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.396330 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.497671 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.497718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.497771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nsfh\" (UniqueName: \"kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.498351 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.498412 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.529779 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nsfh\" (UniqueName: \"kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh\") pod \"redhat-marketplace-ldwlm\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:23 crc kubenswrapper[4959]: I1003 14:27:23.610471 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:24 crc kubenswrapper[4959]: I1003 14:27:24.075780 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:24 crc kubenswrapper[4959]: I1003 14:27:24.237534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerStarted","Data":"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528"} Oct 03 14:27:24 crc kubenswrapper[4959]: I1003 14:27:24.237886 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerStarted","Data":"7f48de3627b5b40367b45f3f123a6a6d20b726c22f19102b5b87b2cadb030a9d"} Oct 03 14:27:25 crc kubenswrapper[4959]: I1003 14:27:25.244516 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c969d32-be33-4c26-8f3e-09489f126c03" containerID="51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528" exitCode=0 Oct 03 14:27:25 crc kubenswrapper[4959]: I1003 14:27:25.244564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerDied","Data":"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528"} Oct 03 14:27:25 crc kubenswrapper[4959]: I1003 14:27:25.246416 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:27:27 crc kubenswrapper[4959]: I1003 14:27:27.262432 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c969d32-be33-4c26-8f3e-09489f126c03" containerID="25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1" exitCode=0 Oct 03 14:27:27 crc kubenswrapper[4959]: I1003 14:27:27.262769 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerDied","Data":"25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1"} Oct 03 14:27:28 crc kubenswrapper[4959]: I1003 14:27:28.273171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerStarted","Data":"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9"} Oct 03 14:27:28 crc kubenswrapper[4959]: I1003 14:27:28.288704 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ldwlm" podStartSLOduration=2.770433923 podStartE2EDuration="5.288683815s" podCreationTimestamp="2025-10-03 14:27:23 +0000 UTC" firstStartedPulling="2025-10-03 14:27:25.245825857 +0000 UTC m=+3414.449169274" lastFinishedPulling="2025-10-03 14:27:27.764075749 +0000 UTC m=+3416.967419166" observedRunningTime="2025-10-03 14:27:28.286325277 +0000 UTC m=+3417.489668704" watchObservedRunningTime="2025-10-03 14:27:28.288683815 +0000 UTC m=+3417.492027232" Oct 03 14:27:33 crc kubenswrapper[4959]: I1003 14:27:33.611227 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:33 crc kubenswrapper[4959]: I1003 14:27:33.611502 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:33 crc kubenswrapper[4959]: I1003 14:27:33.667322 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:34 crc kubenswrapper[4959]: I1003 14:27:34.355563 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:34 crc kubenswrapper[4959]: I1003 14:27:34.407956 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.044350 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.044449 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.044495 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.045098 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.045168 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834" gracePeriod=600 Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.337062 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834" exitCode=0 Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.337142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834"} Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.337697 4959 scope.go:117] "RemoveContainer" containerID="5ed97c07a43c4ba59a85aa114fd20b2444cc15448aa3510e90d7cb2bf9c18c24" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.337890 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ldwlm" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="registry-server" containerID="cri-o://6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9" gracePeriod=2 Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.748000 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.884667 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nsfh\" (UniqueName: \"kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh\") pod \"2c969d32-be33-4c26-8f3e-09489f126c03\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.884770 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content\") pod \"2c969d32-be33-4c26-8f3e-09489f126c03\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.884789 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities\") pod \"2c969d32-be33-4c26-8f3e-09489f126c03\" (UID: \"2c969d32-be33-4c26-8f3e-09489f126c03\") " Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.886464 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities" (OuterVolumeSpecName: "utilities") pod "2c969d32-be33-4c26-8f3e-09489f126c03" (UID: "2c969d32-be33-4c26-8f3e-09489f126c03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.890504 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh" (OuterVolumeSpecName: "kube-api-access-4nsfh") pod "2c969d32-be33-4c26-8f3e-09489f126c03" (UID: "2c969d32-be33-4c26-8f3e-09489f126c03"). InnerVolumeSpecName "kube-api-access-4nsfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.898900 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c969d32-be33-4c26-8f3e-09489f126c03" (UID: "2c969d32-be33-4c26-8f3e-09489f126c03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.986722 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.986945 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c969d32-be33-4c26-8f3e-09489f126c03-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:27:36 crc kubenswrapper[4959]: I1003 14:27:36.987038 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nsfh\" (UniqueName: \"kubernetes.io/projected/2c969d32-be33-4c26-8f3e-09489f126c03-kube-api-access-4nsfh\") on node \"crc\" DevicePath \"\"" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.352604 4959 generic.go:334] "Generic (PLEG): container finished" podID="2c969d32-be33-4c26-8f3e-09489f126c03" containerID="6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9" exitCode=0 Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.352653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerDied","Data":"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9"} Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.352708 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldwlm" event={"ID":"2c969d32-be33-4c26-8f3e-09489f126c03","Type":"ContainerDied","Data":"7f48de3627b5b40367b45f3f123a6a6d20b726c22f19102b5b87b2cadb030a9d"} Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.352730 4959 scope.go:117] "RemoveContainer" containerID="6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.353717 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldwlm" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.355270 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5"} Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.373716 4959 scope.go:117] "RemoveContainer" containerID="25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.387973 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.396152 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldwlm"] Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.409115 4959 scope.go:117] "RemoveContainer" containerID="51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.427849 4959 scope.go:117] "RemoveContainer" containerID="6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9" Oct 03 14:27:37 crc kubenswrapper[4959]: E1003 14:27:37.428245 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9\": container with ID starting with 6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9 not found: ID does not exist" containerID="6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.428295 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9"} err="failed to get container status \"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9\": rpc error: code = NotFound desc = could not find container \"6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9\": container with ID starting with 6c0fbdf90d98a37a5e36be3127fe2a46a6e2a4d2dd604d6f8abac35704a744e9 not found: ID does not exist" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.428356 4959 scope.go:117] "RemoveContainer" containerID="25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1" Oct 03 14:27:37 crc kubenswrapper[4959]: E1003 14:27:37.429098 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1\": container with ID starting with 25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1 not found: ID does not exist" containerID="25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.429243 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1"} err="failed to get container status \"25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1\": rpc error: code = NotFound desc = could not find container \"25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1\": container with ID starting with 25e2a5d0e81b924a6b69063340f5d9051b4d90e324e5222ae7538fff4aba15c1 not found: ID does not exist" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.429349 4959 scope.go:117] "RemoveContainer" containerID="51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528" Oct 03 14:27:37 crc kubenswrapper[4959]: E1003 14:27:37.429924 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528\": container with ID starting with 51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528 not found: ID does not exist" containerID="51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.430046 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528"} err="failed to get container status \"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528\": rpc error: code = NotFound desc = could not find container \"51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528\": container with ID starting with 51c9338b248d3e5d6c6e79a07a712ccae57df9925d79cbacb7543fc0fa824528 not found: ID does not exist" Oct 03 14:27:37 crc kubenswrapper[4959]: I1003 14:27:37.698030 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" path="/var/lib/kubelet/pods/2c969d32-be33-4c26-8f3e-09489f126c03/volumes" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.489797 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:28 crc kubenswrapper[4959]: E1003 14:28:28.490873 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="registry-server" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.490896 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="registry-server" Oct 03 14:28:28 crc kubenswrapper[4959]: E1003 14:28:28.490927 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="extract-utilities" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.490940 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="extract-utilities" Oct 03 14:28:28 crc kubenswrapper[4959]: E1003 14:28:28.490989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="extract-content" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.491003 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="extract-content" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.491336 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c969d32-be33-4c26-8f3e-09489f126c03" containerName="registry-server" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.494931 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.497740 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.625753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.625882 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6fj5\" (UniqueName: \"kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.625948 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.727847 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.727938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.728030 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6fj5\" (UniqueName: \"kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.728766 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.729145 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.748634 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6fj5\" (UniqueName: \"kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5\") pod \"redhat-operators-szwnc\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:28 crc kubenswrapper[4959]: I1003 14:28:28.837092 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:29 crc kubenswrapper[4959]: I1003 14:28:29.347844 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:29 crc kubenswrapper[4959]: I1003 14:28:29.740039 4959 generic.go:334] "Generic (PLEG): container finished" podID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerID="6393e597eafce92a442477d43b3f78021a573473a0443ee549a18a1cb1c40e1e" exitCode=0 Oct 03 14:28:29 crc kubenswrapper[4959]: I1003 14:28:29.740247 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerDied","Data":"6393e597eafce92a442477d43b3f78021a573473a0443ee549a18a1cb1c40e1e"} Oct 03 14:28:29 crc kubenswrapper[4959]: I1003 14:28:29.741324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerStarted","Data":"49f6630a48009ce8580a543ffa26cf4b12b19006e907ca79331b60cb7541b32e"} Oct 03 14:28:30 crc kubenswrapper[4959]: I1003 14:28:30.749876 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerStarted","Data":"1d80b56df8669a5a2b0474823e8760f8cd1df464781ea1686d694e06a1e267f6"} Oct 03 14:28:31 crc kubenswrapper[4959]: I1003 14:28:31.758154 4959 generic.go:334] "Generic (PLEG): container finished" podID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerID="1d80b56df8669a5a2b0474823e8760f8cd1df464781ea1686d694e06a1e267f6" exitCode=0 Oct 03 14:28:31 crc kubenswrapper[4959]: I1003 14:28:31.758331 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerDied","Data":"1d80b56df8669a5a2b0474823e8760f8cd1df464781ea1686d694e06a1e267f6"} Oct 03 14:28:32 crc kubenswrapper[4959]: I1003 14:28:32.774359 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerStarted","Data":"589981bcd6c68731ac6afc028bca672f235bf37ffc5b82f507ad1fb50f233270"} Oct 03 14:28:32 crc kubenswrapper[4959]: I1003 14:28:32.789797 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-szwnc" podStartSLOduration=2.101658339 podStartE2EDuration="4.789776677s" podCreationTimestamp="2025-10-03 14:28:28 +0000 UTC" firstStartedPulling="2025-10-03 14:28:29.741894215 +0000 UTC m=+3478.945237632" lastFinishedPulling="2025-10-03 14:28:32.430012553 +0000 UTC m=+3481.633355970" observedRunningTime="2025-10-03 14:28:32.787597214 +0000 UTC m=+3481.990940641" watchObservedRunningTime="2025-10-03 14:28:32.789776677 +0000 UTC m=+3481.993120094" Oct 03 14:28:38 crc kubenswrapper[4959]: I1003 14:28:38.839058 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:38 crc kubenswrapper[4959]: I1003 14:28:38.839907 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:38 crc kubenswrapper[4959]: I1003 14:28:38.900687 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:39 crc kubenswrapper[4959]: I1003 14:28:39.862786 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:39 crc kubenswrapper[4959]: I1003 14:28:39.908935 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:41 crc kubenswrapper[4959]: I1003 14:28:41.844175 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-szwnc" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="registry-server" containerID="cri-o://589981bcd6c68731ac6afc028bca672f235bf37ffc5b82f507ad1fb50f233270" gracePeriod=2 Oct 03 14:28:42 crc kubenswrapper[4959]: I1003 14:28:42.851929 4959 generic.go:334] "Generic (PLEG): container finished" podID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerID="589981bcd6c68731ac6afc028bca672f235bf37ffc5b82f507ad1fb50f233270" exitCode=0 Oct 03 14:28:42 crc kubenswrapper[4959]: I1003 14:28:42.852006 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerDied","Data":"589981bcd6c68731ac6afc028bca672f235bf37ffc5b82f507ad1fb50f233270"} Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.413090 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.534389 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities\") pod \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.534545 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content\") pod \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.534624 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6fj5\" (UniqueName: \"kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5\") pod \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\" (UID: \"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d\") " Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.535575 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities" (OuterVolumeSpecName: "utilities") pod "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" (UID: "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.541011 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5" (OuterVolumeSpecName: "kube-api-access-q6fj5") pod "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" (UID: "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d"). InnerVolumeSpecName "kube-api-access-q6fj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.618935 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" (UID: "e2e1a970-ca77-43c6-8592-7efe3ebf1b2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.636069 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.636094 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6fj5\" (UniqueName: \"kubernetes.io/projected/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-kube-api-access-q6fj5\") on node \"crc\" DevicePath \"\"" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.636106 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.869218 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-szwnc" event={"ID":"e2e1a970-ca77-43c6-8592-7efe3ebf1b2d","Type":"ContainerDied","Data":"49f6630a48009ce8580a543ffa26cf4b12b19006e907ca79331b60cb7541b32e"} Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.869287 4959 scope.go:117] "RemoveContainer" containerID="589981bcd6c68731ac6afc028bca672f235bf37ffc5b82f507ad1fb50f233270" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.869299 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-szwnc" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.901049 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.905668 4959 scope.go:117] "RemoveContainer" containerID="1d80b56df8669a5a2b0474823e8760f8cd1df464781ea1686d694e06a1e267f6" Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.906854 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-szwnc"] Oct 03 14:28:43 crc kubenswrapper[4959]: I1003 14:28:43.925709 4959 scope.go:117] "RemoveContainer" containerID="6393e597eafce92a442477d43b3f78021a573473a0443ee549a18a1cb1c40e1e" Oct 03 14:28:45 crc kubenswrapper[4959]: I1003 14:28:45.697107 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" path="/var/lib/kubelet/pods/e2e1a970-ca77-43c6-8592-7efe3ebf1b2d/volumes" Oct 03 14:29:36 crc kubenswrapper[4959]: I1003 14:29:36.044678 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:29:36 crc kubenswrapper[4959]: I1003 14:29:36.045300 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.182663 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt"] Oct 03 14:30:00 crc kubenswrapper[4959]: E1003 14:30:00.183550 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="registry-server" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.183567 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="registry-server" Oct 03 14:30:00 crc kubenswrapper[4959]: E1003 14:30:00.183582 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="extract-utilities" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.183590 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="extract-utilities" Oct 03 14:30:00 crc kubenswrapper[4959]: E1003 14:30:00.183608 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="extract-content" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.183616 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="extract-content" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.183789 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2e1a970-ca77-43c6-8592-7efe3ebf1b2d" containerName="registry-server" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.184399 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.186651 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.192506 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.194688 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt"] Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.244066 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwst7\" (UniqueName: \"kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.244110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.244167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.345467 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.345568 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwst7\" (UniqueName: \"kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.345592 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.346572 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.356798 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.362434 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwst7\" (UniqueName: \"kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7\") pod \"collect-profiles-29325030-pt8dt\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.545679 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:00 crc kubenswrapper[4959]: I1003 14:30:00.935567 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt"] Oct 03 14:30:01 crc kubenswrapper[4959]: I1003 14:30:01.446080 4959 generic.go:334] "Generic (PLEG): container finished" podID="cb063e10-6487-48b0-84c9-258187deb68b" containerID="f17fdd0fa725c3ba22091e35e6dd8f60c47ccebcabf002a913e97455899d93ce" exitCode=0 Oct 03 14:30:01 crc kubenswrapper[4959]: I1003 14:30:01.446234 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" event={"ID":"cb063e10-6487-48b0-84c9-258187deb68b","Type":"ContainerDied","Data":"f17fdd0fa725c3ba22091e35e6dd8f60c47ccebcabf002a913e97455899d93ce"} Oct 03 14:30:01 crc kubenswrapper[4959]: I1003 14:30:01.446534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" event={"ID":"cb063e10-6487-48b0-84c9-258187deb68b","Type":"ContainerStarted","Data":"31c4a2437484767da8db41bd7d198bf3e9a7eb7a7607d69593af27bc4db42ab2"} Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.716861 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.884158 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume\") pod \"cb063e10-6487-48b0-84c9-258187deb68b\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.884262 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume\") pod \"cb063e10-6487-48b0-84c9-258187deb68b\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.884299 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwst7\" (UniqueName: \"kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7\") pod \"cb063e10-6487-48b0-84c9-258187deb68b\" (UID: \"cb063e10-6487-48b0-84c9-258187deb68b\") " Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.885048 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume" (OuterVolumeSpecName: "config-volume") pod "cb063e10-6487-48b0-84c9-258187deb68b" (UID: "cb063e10-6487-48b0-84c9-258187deb68b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.889001 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7" (OuterVolumeSpecName: "kube-api-access-rwst7") pod "cb063e10-6487-48b0-84c9-258187deb68b" (UID: "cb063e10-6487-48b0-84c9-258187deb68b"). InnerVolumeSpecName "kube-api-access-rwst7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.889136 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cb063e10-6487-48b0-84c9-258187deb68b" (UID: "cb063e10-6487-48b0-84c9-258187deb68b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.985977 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb063e10-6487-48b0-84c9-258187deb68b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.986008 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb063e10-6487-48b0-84c9-258187deb68b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:30:02 crc kubenswrapper[4959]: I1003 14:30:02.986018 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwst7\" (UniqueName: \"kubernetes.io/projected/cb063e10-6487-48b0-84c9-258187deb68b-kube-api-access-rwst7\") on node \"crc\" DevicePath \"\"" Oct 03 14:30:03 crc kubenswrapper[4959]: I1003 14:30:03.462597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" event={"ID":"cb063e10-6487-48b0-84c9-258187deb68b","Type":"ContainerDied","Data":"31c4a2437484767da8db41bd7d198bf3e9a7eb7a7607d69593af27bc4db42ab2"} Oct 03 14:30:03 crc kubenswrapper[4959]: I1003 14:30:03.462639 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c4a2437484767da8db41bd7d198bf3e9a7eb7a7607d69593af27bc4db42ab2" Oct 03 14:30:03 crc kubenswrapper[4959]: I1003 14:30:03.462650 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt" Oct 03 14:30:03 crc kubenswrapper[4959]: I1003 14:30:03.780127 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm"] Oct 03 14:30:03 crc kubenswrapper[4959]: I1003 14:30:03.787514 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-mdznm"] Oct 03 14:30:05 crc kubenswrapper[4959]: I1003 14:30:05.693623 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e9c3104-76bd-4252-9901-37e9d6d14632" path="/var/lib/kubelet/pods/1e9c3104-76bd-4252-9901-37e9d6d14632/volumes" Oct 03 14:30:06 crc kubenswrapper[4959]: I1003 14:30:06.043995 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:30:06 crc kubenswrapper[4959]: I1003 14:30:06.044054 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.044473 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.045275 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.045378 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.046470 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.046604 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" gracePeriod=600 Oct 03 14:30:36 crc kubenswrapper[4959]: E1003 14:30:36.167655 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.707707 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" exitCode=0 Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.707774 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5"} Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.708016 4959 scope.go:117] "RemoveContainer" containerID="56a4baca47101c13017058ad80b52166f05dd5e74ce01b602c2bb0d1711b1834" Oct 03 14:30:36 crc kubenswrapper[4959]: I1003 14:30:36.708628 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:30:36 crc kubenswrapper[4959]: E1003 14:30:36.708857 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:30:49 crc kubenswrapper[4959]: I1003 14:30:49.685985 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:30:49 crc kubenswrapper[4959]: E1003 14:30:49.686791 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:31:00 crc kubenswrapper[4959]: I1003 14:31:00.685672 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:31:00 crc kubenswrapper[4959]: E1003 14:31:00.686677 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:31:04 crc kubenswrapper[4959]: I1003 14:31:04.279520 4959 scope.go:117] "RemoveContainer" containerID="2aeef83766ab49120fc33f5fe8c4e31f870ab2e9e8bcffe8be170d4434c8caff" Oct 03 14:31:15 crc kubenswrapper[4959]: I1003 14:31:15.685552 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:31:15 crc kubenswrapper[4959]: E1003 14:31:15.686600 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:31:29 crc kubenswrapper[4959]: I1003 14:31:29.685868 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:31:29 crc kubenswrapper[4959]: E1003 14:31:29.686578 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:31:44 crc kubenswrapper[4959]: I1003 14:31:44.685368 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:31:44 crc kubenswrapper[4959]: E1003 14:31:44.686562 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:31:56 crc kubenswrapper[4959]: I1003 14:31:56.685571 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:31:56 crc kubenswrapper[4959]: E1003 14:31:56.686171 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:32:11 crc kubenswrapper[4959]: I1003 14:32:11.690093 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:32:11 crc kubenswrapper[4959]: E1003 14:32:11.691115 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.685966 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:32:24 crc kubenswrapper[4959]: E1003 14:32:24.686809 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.906169 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:24 crc kubenswrapper[4959]: E1003 14:32:24.906721 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb063e10-6487-48b0-84c9-258187deb68b" containerName="collect-profiles" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.906753 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb063e10-6487-48b0-84c9-258187deb68b" containerName="collect-profiles" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.907036 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb063e10-6487-48b0-84c9-258187deb68b" containerName="collect-profiles" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.908949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:24 crc kubenswrapper[4959]: I1003 14:32:24.917038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.043464 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.043551 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjsqs\" (UniqueName: \"kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.043737 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.144958 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.145005 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjsqs\" (UniqueName: \"kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.145095 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.145631 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.145661 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.169376 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjsqs\" (UniqueName: \"kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs\") pod \"community-operators-95jvw\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.242151 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.485722 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:25 crc kubenswrapper[4959]: I1003 14:32:25.544605 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerStarted","Data":"e1b47719aa5c09cdc60209fca3777be59ebaaf48e3b9a44693166b8baac50f5d"} Oct 03 14:32:26 crc kubenswrapper[4959]: I1003 14:32:26.579565 4959 generic.go:334] "Generic (PLEG): container finished" podID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerID="d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d" exitCode=0 Oct 03 14:32:26 crc kubenswrapper[4959]: I1003 14:32:26.579617 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerDied","Data":"d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d"} Oct 03 14:32:26 crc kubenswrapper[4959]: I1003 14:32:26.584332 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.183378 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.185773 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.191768 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.318101 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.318222 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.318626 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q99mr\" (UniqueName: \"kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.419764 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.420380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q99mr\" (UniqueName: \"kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.420303 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.420859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.421243 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.442853 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q99mr\" (UniqueName: \"kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr\") pod \"certified-operators-2dl76\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.516604 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.639046 4959 generic.go:334] "Generic (PLEG): container finished" podID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerID="fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1" exitCode=0 Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.639380 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerDied","Data":"fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1"} Oct 03 14:32:30 crc kubenswrapper[4959]: I1003 14:32:30.989964 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:31 crc kubenswrapper[4959]: I1003 14:32:31.650816 4959 generic.go:334] "Generic (PLEG): container finished" podID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerID="0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c" exitCode=0 Oct 03 14:32:31 crc kubenswrapper[4959]: I1003 14:32:31.650887 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerDied","Data":"0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c"} Oct 03 14:32:31 crc kubenswrapper[4959]: I1003 14:32:31.651868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerStarted","Data":"ad9b407a9e6aa0f8cd4358f3b4b2565bcadafceda6e6803f6509d6149b4c9fe6"} Oct 03 14:32:32 crc kubenswrapper[4959]: I1003 14:32:32.663030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerStarted","Data":"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab"} Oct 03 14:32:32 crc kubenswrapper[4959]: I1003 14:32:32.679536 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-95jvw" podStartSLOduration=3.612238818 podStartE2EDuration="8.679521295s" podCreationTimestamp="2025-10-03 14:32:24 +0000 UTC" firstStartedPulling="2025-10-03 14:32:26.582522213 +0000 UTC m=+3715.785865640" lastFinishedPulling="2025-10-03 14:32:31.6498047 +0000 UTC m=+3720.853148117" observedRunningTime="2025-10-03 14:32:32.677969067 +0000 UTC m=+3721.881312484" watchObservedRunningTime="2025-10-03 14:32:32.679521295 +0000 UTC m=+3721.882864712" Oct 03 14:32:33 crc kubenswrapper[4959]: I1003 14:32:33.671144 4959 generic.go:334] "Generic (PLEG): container finished" podID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerID="42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752" exitCode=0 Oct 03 14:32:33 crc kubenswrapper[4959]: I1003 14:32:33.671710 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerDied","Data":"42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752"} Oct 03 14:32:35 crc kubenswrapper[4959]: I1003 14:32:35.243123 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:35 crc kubenswrapper[4959]: I1003 14:32:35.243529 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:35 crc kubenswrapper[4959]: I1003 14:32:35.300358 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:35 crc kubenswrapper[4959]: I1003 14:32:35.707662 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerStarted","Data":"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21"} Oct 03 14:32:35 crc kubenswrapper[4959]: I1003 14:32:35.721987 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dl76" podStartSLOduration=2.5035160469999997 podStartE2EDuration="5.721969003s" podCreationTimestamp="2025-10-03 14:32:30 +0000 UTC" firstStartedPulling="2025-10-03 14:32:31.654128925 +0000 UTC m=+3720.857472372" lastFinishedPulling="2025-10-03 14:32:34.872581871 +0000 UTC m=+3724.075925328" observedRunningTime="2025-10-03 14:32:35.717266347 +0000 UTC m=+3724.920609784" watchObservedRunningTime="2025-10-03 14:32:35.721969003 +0000 UTC m=+3724.925312420" Oct 03 14:32:39 crc kubenswrapper[4959]: I1003 14:32:39.686005 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:32:39 crc kubenswrapper[4959]: E1003 14:32:39.686710 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:32:40 crc kubenswrapper[4959]: I1003 14:32:40.516771 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:40 crc kubenswrapper[4959]: I1003 14:32:40.516828 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:40 crc kubenswrapper[4959]: I1003 14:32:40.570630 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:40 crc kubenswrapper[4959]: I1003 14:32:40.766859 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:40 crc kubenswrapper[4959]: I1003 14:32:40.814590 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:42 crc kubenswrapper[4959]: I1003 14:32:42.739714 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dl76" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="registry-server" containerID="cri-o://58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21" gracePeriod=2 Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.136406 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.306455 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities\") pod \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.306516 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content\") pod \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.306606 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q99mr\" (UniqueName: \"kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr\") pod \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\" (UID: \"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9\") " Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.308027 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities" (OuterVolumeSpecName: "utilities") pod "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" (UID: "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.311833 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr" (OuterVolumeSpecName: "kube-api-access-q99mr") pod "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" (UID: "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9"). InnerVolumeSpecName "kube-api-access-q99mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.359277 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" (UID: "7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.407939 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.408006 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.408022 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q99mr\" (UniqueName: \"kubernetes.io/projected/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9-kube-api-access-q99mr\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.750333 4959 generic.go:334] "Generic (PLEG): container finished" podID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerID="58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21" exitCode=0 Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.750383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerDied","Data":"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21"} Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.750401 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dl76" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.750419 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dl76" event={"ID":"7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9","Type":"ContainerDied","Data":"ad9b407a9e6aa0f8cd4358f3b4b2565bcadafceda6e6803f6509d6149b4c9fe6"} Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.750445 4959 scope.go:117] "RemoveContainer" containerID="58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.775485 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.780918 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dl76"] Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.781088 4959 scope.go:117] "RemoveContainer" containerID="42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.803998 4959 scope.go:117] "RemoveContainer" containerID="0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.831220 4959 scope.go:117] "RemoveContainer" containerID="58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21" Oct 03 14:32:43 crc kubenswrapper[4959]: E1003 14:32:43.831831 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21\": container with ID starting with 58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21 not found: ID does not exist" containerID="58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.831859 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21"} err="failed to get container status \"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21\": rpc error: code = NotFound desc = could not find container \"58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21\": container with ID starting with 58ab0ad7bbf64d50063e8ef93e91112d27aa578e62355316e68050957f6a4f21 not found: ID does not exist" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.831879 4959 scope.go:117] "RemoveContainer" containerID="42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752" Oct 03 14:32:43 crc kubenswrapper[4959]: E1003 14:32:43.832141 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752\": container with ID starting with 42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752 not found: ID does not exist" containerID="42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.832166 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752"} err="failed to get container status \"42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752\": rpc error: code = NotFound desc = could not find container \"42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752\": container with ID starting with 42ce4c3e1039c1c729ec36d10bb0af33826f03e9948ea3a280cd48a2959cc752 not found: ID does not exist" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.832185 4959 scope.go:117] "RemoveContainer" containerID="0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c" Oct 03 14:32:43 crc kubenswrapper[4959]: E1003 14:32:43.832570 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c\": container with ID starting with 0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c not found: ID does not exist" containerID="0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c" Oct 03 14:32:43 crc kubenswrapper[4959]: I1003 14:32:43.832677 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c"} err="failed to get container status \"0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c\": rpc error: code = NotFound desc = could not find container \"0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c\": container with ID starting with 0dcf8e9d6a2b3f64d5fc012447e05d6969aa2b25ed9158cb6493ce836ffd6a7c not found: ID does not exist" Oct 03 14:32:45 crc kubenswrapper[4959]: I1003 14:32:45.303920 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:45 crc kubenswrapper[4959]: I1003 14:32:45.695450 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" path="/var/lib/kubelet/pods/7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9/volumes" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.208308 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.208603 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-95jvw" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="registry-server" containerID="cri-o://ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab" gracePeriod=2 Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.614025 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.755589 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjsqs\" (UniqueName: \"kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs\") pod \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.755648 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content\") pod \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.755777 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities\") pod \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\" (UID: \"52d49bbe-ab89-4bbd-8f46-157c4c20805c\") " Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.756667 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities" (OuterVolumeSpecName: "utilities") pod "52d49bbe-ab89-4bbd-8f46-157c4c20805c" (UID: "52d49bbe-ab89-4bbd-8f46-157c4c20805c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.761338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs" (OuterVolumeSpecName: "kube-api-access-bjsqs") pod "52d49bbe-ab89-4bbd-8f46-157c4c20805c" (UID: "52d49bbe-ab89-4bbd-8f46-157c4c20805c"). InnerVolumeSpecName "kube-api-access-bjsqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.773639 4959 generic.go:334] "Generic (PLEG): container finished" podID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerID="ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab" exitCode=0 Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.773744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerDied","Data":"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab"} Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.773779 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95jvw" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.773811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95jvw" event={"ID":"52d49bbe-ab89-4bbd-8f46-157c4c20805c","Type":"ContainerDied","Data":"e1b47719aa5c09cdc60209fca3777be59ebaaf48e3b9a44693166b8baac50f5d"} Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.773835 4959 scope.go:117] "RemoveContainer" containerID="ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.801810 4959 scope.go:117] "RemoveContainer" containerID="fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.806586 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52d49bbe-ab89-4bbd-8f46-157c4c20805c" (UID: "52d49bbe-ab89-4bbd-8f46-157c4c20805c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.833399 4959 scope.go:117] "RemoveContainer" containerID="d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.847910 4959 scope.go:117] "RemoveContainer" containerID="ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab" Oct 03 14:32:46 crc kubenswrapper[4959]: E1003 14:32:46.848459 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab\": container with ID starting with ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab not found: ID does not exist" containerID="ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.848496 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab"} err="failed to get container status \"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab\": rpc error: code = NotFound desc = could not find container \"ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab\": container with ID starting with ad60bfdf74965a9300d0e738aec7cb7b92e2d7c834f384f2e100d486b253f3ab not found: ID does not exist" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.848521 4959 scope.go:117] "RemoveContainer" containerID="fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1" Oct 03 14:32:46 crc kubenswrapper[4959]: E1003 14:32:46.848831 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1\": container with ID starting with fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1 not found: ID does not exist" containerID="fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.848871 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1"} err="failed to get container status \"fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1\": rpc error: code = NotFound desc = could not find container \"fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1\": container with ID starting with fe9a54b0bfd199b451ea655c8f15747a529436353b85fb9c80a148980cfb17c1 not found: ID does not exist" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.848899 4959 scope.go:117] "RemoveContainer" containerID="d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d" Oct 03 14:32:46 crc kubenswrapper[4959]: E1003 14:32:46.849253 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d\": container with ID starting with d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d not found: ID does not exist" containerID="d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.849289 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d"} err="failed to get container status \"d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d\": rpc error: code = NotFound desc = could not find container \"d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d\": container with ID starting with d481b1343e4433e4fa28c226f0e60422dabaf38b6d5d658bdc8b94d4947a011d not found: ID does not exist" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.857339 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.857389 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjsqs\" (UniqueName: \"kubernetes.io/projected/52d49bbe-ab89-4bbd-8f46-157c4c20805c-kube-api-access-bjsqs\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:46 crc kubenswrapper[4959]: I1003 14:32:46.857406 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52d49bbe-ab89-4bbd-8f46-157c4c20805c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:32:47 crc kubenswrapper[4959]: I1003 14:32:47.104074 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:47 crc kubenswrapper[4959]: I1003 14:32:47.111460 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-95jvw"] Oct 03 14:32:47 crc kubenswrapper[4959]: I1003 14:32:47.700034 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" path="/var/lib/kubelet/pods/52d49bbe-ab89-4bbd-8f46-157c4c20805c/volumes" Oct 03 14:32:50 crc kubenswrapper[4959]: I1003 14:32:50.686386 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:32:50 crc kubenswrapper[4959]: E1003 14:32:50.686924 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:33:02 crc kubenswrapper[4959]: I1003 14:33:02.686056 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:33:02 crc kubenswrapper[4959]: E1003 14:33:02.688287 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:33:15 crc kubenswrapper[4959]: I1003 14:33:15.685270 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:33:15 crc kubenswrapper[4959]: E1003 14:33:15.685934 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:33:27 crc kubenswrapper[4959]: I1003 14:33:27.685591 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:33:27 crc kubenswrapper[4959]: E1003 14:33:27.686159 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:33:39 crc kubenswrapper[4959]: I1003 14:33:39.685733 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:33:39 crc kubenswrapper[4959]: E1003 14:33:39.687032 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:33:54 crc kubenswrapper[4959]: I1003 14:33:54.686530 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:33:54 crc kubenswrapper[4959]: E1003 14:33:54.687569 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:34:08 crc kubenswrapper[4959]: I1003 14:34:08.684931 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:34:08 crc kubenswrapper[4959]: E1003 14:34:08.685626 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:34:21 crc kubenswrapper[4959]: I1003 14:34:21.690109 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:34:21 crc kubenswrapper[4959]: E1003 14:34:21.691518 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:34:36 crc kubenswrapper[4959]: I1003 14:34:36.686181 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:34:36 crc kubenswrapper[4959]: E1003 14:34:36.687327 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:34:49 crc kubenswrapper[4959]: I1003 14:34:49.685942 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:34:49 crc kubenswrapper[4959]: E1003 14:34:49.688378 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:35:01 crc kubenswrapper[4959]: I1003 14:35:01.696337 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:35:01 crc kubenswrapper[4959]: E1003 14:35:01.697496 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:35:16 crc kubenswrapper[4959]: I1003 14:35:16.685384 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:35:16 crc kubenswrapper[4959]: E1003 14:35:16.686055 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:35:28 crc kubenswrapper[4959]: I1003 14:35:28.685844 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:35:28 crc kubenswrapper[4959]: E1003 14:35:28.687232 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:35:41 crc kubenswrapper[4959]: I1003 14:35:41.691550 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:35:42 crc kubenswrapper[4959]: I1003 14:35:42.132373 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6"} Oct 03 14:38:06 crc kubenswrapper[4959]: I1003 14:38:06.045114 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:38:06 crc kubenswrapper[4959]: I1003 14:38:06.046887 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.515547 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516090 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="extract-content" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516104 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="extract-content" Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516123 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="extract-utilities" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516130 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="extract-utilities" Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516142 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516148 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516161 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516169 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516244 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="extract-utilities" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516252 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="extract-utilities" Oct 03 14:38:09 crc kubenswrapper[4959]: E1003 14:38:09.516261 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="extract-content" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516268 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="extract-content" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516405 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a1fd8dc-6633-4c4a-86ee-e5d53ca851e9" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.516419 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d49bbe-ab89-4bbd-8f46-157c4c20805c" containerName="registry-server" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.517474 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.532004 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.586778 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.586906 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz5h9\" (UniqueName: \"kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.586937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.705837 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.706240 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz5h9\" (UniqueName: \"kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.706324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.706881 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.706908 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.743294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz5h9\" (UniqueName: \"kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9\") pod \"redhat-marketplace-xbqnr\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:09 crc kubenswrapper[4959]: I1003 14:38:09.840886 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:10 crc kubenswrapper[4959]: I1003 14:38:10.249458 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:10 crc kubenswrapper[4959]: I1003 14:38:10.369052 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerStarted","Data":"0549e5df0df8a0448b6988f09dd42c71357cbae5a487e7eeedd33f70de4446ca"} Oct 03 14:38:11 crc kubenswrapper[4959]: I1003 14:38:11.378523 4959 generic.go:334] "Generic (PLEG): container finished" podID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerID="f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61" exitCode=0 Oct 03 14:38:11 crc kubenswrapper[4959]: I1003 14:38:11.378627 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerDied","Data":"f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61"} Oct 03 14:38:11 crc kubenswrapper[4959]: I1003 14:38:11.380858 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:38:13 crc kubenswrapper[4959]: I1003 14:38:13.397153 4959 generic.go:334] "Generic (PLEG): container finished" podID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerID="b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9" exitCode=0 Oct 03 14:38:13 crc kubenswrapper[4959]: I1003 14:38:13.398394 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerDied","Data":"b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9"} Oct 03 14:38:14 crc kubenswrapper[4959]: I1003 14:38:14.407860 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerStarted","Data":"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e"} Oct 03 14:38:14 crc kubenswrapper[4959]: I1003 14:38:14.438131 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xbqnr" podStartSLOduration=2.981011566 podStartE2EDuration="5.438105608s" podCreationTimestamp="2025-10-03 14:38:09 +0000 UTC" firstStartedPulling="2025-10-03 14:38:11.380639111 +0000 UTC m=+4060.583982528" lastFinishedPulling="2025-10-03 14:38:13.837733153 +0000 UTC m=+4063.041076570" observedRunningTime="2025-10-03 14:38:14.422487604 +0000 UTC m=+4063.625831031" watchObservedRunningTime="2025-10-03 14:38:14.438105608 +0000 UTC m=+4063.641449025" Oct 03 14:38:19 crc kubenswrapper[4959]: I1003 14:38:19.841390 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:19 crc kubenswrapper[4959]: I1003 14:38:19.841950 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:19 crc kubenswrapper[4959]: I1003 14:38:19.886317 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:20 crc kubenswrapper[4959]: I1003 14:38:20.541535 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:20 crc kubenswrapper[4959]: I1003 14:38:20.604661 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.481468 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xbqnr" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="registry-server" containerID="cri-o://8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e" gracePeriod=2 Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.847249 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.907665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities\") pod \"0703572c-8efc-4c98-a8f4-15f94a40ea01\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.907707 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content\") pod \"0703572c-8efc-4c98-a8f4-15f94a40ea01\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.907737 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz5h9\" (UniqueName: \"kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9\") pod \"0703572c-8efc-4c98-a8f4-15f94a40ea01\" (UID: \"0703572c-8efc-4c98-a8f4-15f94a40ea01\") " Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.908548 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities" (OuterVolumeSpecName: "utilities") pod "0703572c-8efc-4c98-a8f4-15f94a40ea01" (UID: "0703572c-8efc-4c98-a8f4-15f94a40ea01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.912834 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9" (OuterVolumeSpecName: "kube-api-access-hz5h9") pod "0703572c-8efc-4c98-a8f4-15f94a40ea01" (UID: "0703572c-8efc-4c98-a8f4-15f94a40ea01"). InnerVolumeSpecName "kube-api-access-hz5h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:38:22 crc kubenswrapper[4959]: I1003 14:38:22.920290 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0703572c-8efc-4c98-a8f4-15f94a40ea01" (UID: "0703572c-8efc-4c98-a8f4-15f94a40ea01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.009514 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.009551 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz5h9\" (UniqueName: \"kubernetes.io/projected/0703572c-8efc-4c98-a8f4-15f94a40ea01-kube-api-access-hz5h9\") on node \"crc\" DevicePath \"\"" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.009572 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0703572c-8efc-4c98-a8f4-15f94a40ea01-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.494593 4959 generic.go:334] "Generic (PLEG): container finished" podID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerID="8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e" exitCode=0 Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.494639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerDied","Data":"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e"} Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.494672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xbqnr" event={"ID":"0703572c-8efc-4c98-a8f4-15f94a40ea01","Type":"ContainerDied","Data":"0549e5df0df8a0448b6988f09dd42c71357cbae5a487e7eeedd33f70de4446ca"} Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.494696 4959 scope.go:117] "RemoveContainer" containerID="8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.494702 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xbqnr" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.512472 4959 scope.go:117] "RemoveContainer" containerID="b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.541603 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.547110 4959 scope.go:117] "RemoveContainer" containerID="f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.548125 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xbqnr"] Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.581369 4959 scope.go:117] "RemoveContainer" containerID="8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e" Oct 03 14:38:23 crc kubenswrapper[4959]: E1003 14:38:23.581742 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e\": container with ID starting with 8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e not found: ID does not exist" containerID="8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.581785 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e"} err="failed to get container status \"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e\": rpc error: code = NotFound desc = could not find container \"8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e\": container with ID starting with 8197a1d903806499d45078cedac38aca98548af42b226dbb4c9fd77588a02e5e not found: ID does not exist" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.581816 4959 scope.go:117] "RemoveContainer" containerID="b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9" Oct 03 14:38:23 crc kubenswrapper[4959]: E1003 14:38:23.583550 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9\": container with ID starting with b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9 not found: ID does not exist" containerID="b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.583623 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9"} err="failed to get container status \"b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9\": rpc error: code = NotFound desc = could not find container \"b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9\": container with ID starting with b0109a3e5b356b11d225b62fc9074ddf06c02468f43b05eb5d2ffccf99eb6db9 not found: ID does not exist" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.583723 4959 scope.go:117] "RemoveContainer" containerID="f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61" Oct 03 14:38:23 crc kubenswrapper[4959]: E1003 14:38:23.584183 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61\": container with ID starting with f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61 not found: ID does not exist" containerID="f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.584236 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61"} err="failed to get container status \"f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61\": rpc error: code = NotFound desc = could not find container \"f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61\": container with ID starting with f12b0571922803398b5be25cc09400527cd8bfa05a47f2324b4d4e4555b7ed61 not found: ID does not exist" Oct 03 14:38:23 crc kubenswrapper[4959]: I1003 14:38:23.713598 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" path="/var/lib/kubelet/pods/0703572c-8efc-4c98-a8f4-15f94a40ea01/volumes" Oct 03 14:38:36 crc kubenswrapper[4959]: I1003 14:38:36.045206 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:38:36 crc kubenswrapper[4959]: I1003 14:38:36.046098 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.637379 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:38:49 crc kubenswrapper[4959]: E1003 14:38:49.640039 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="registry-server" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.640186 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="registry-server" Oct 03 14:38:49 crc kubenswrapper[4959]: E1003 14:38:49.640357 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="extract-content" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.640593 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="extract-content" Oct 03 14:38:49 crc kubenswrapper[4959]: E1003 14:38:49.640710 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="extract-utilities" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.640810 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="extract-utilities" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.641131 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0703572c-8efc-4c98-a8f4-15f94a40ea01" containerName="registry-server" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.642896 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.646537 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.721235 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.721356 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tt4l\" (UniqueName: \"kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.721392 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.822889 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.822996 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tt4l\" (UniqueName: \"kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.823033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.823463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.823706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:49 crc kubenswrapper[4959]: I1003 14:38:49.844785 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tt4l\" (UniqueName: \"kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l\") pod \"redhat-operators-k6p82\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:50 crc kubenswrapper[4959]: I1003 14:38:50.021686 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:38:50 crc kubenswrapper[4959]: I1003 14:38:50.430027 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:38:50 crc kubenswrapper[4959]: I1003 14:38:50.710151 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerStarted","Data":"a7ab56db71bb484ebe92d04ea000056108884785a7da4c1f1b8da5ab5222db02"} Oct 03 14:38:51 crc kubenswrapper[4959]: I1003 14:38:51.724338 4959 generic.go:334] "Generic (PLEG): container finished" podID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerID="db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d" exitCode=0 Oct 03 14:38:51 crc kubenswrapper[4959]: I1003 14:38:51.724385 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerDied","Data":"db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d"} Oct 03 14:38:53 crc kubenswrapper[4959]: I1003 14:38:53.742033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerStarted","Data":"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b"} Oct 03 14:38:54 crc kubenswrapper[4959]: I1003 14:38:54.750568 4959 generic.go:334] "Generic (PLEG): container finished" podID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerID="2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b" exitCode=0 Oct 03 14:38:54 crc kubenswrapper[4959]: I1003 14:38:54.750619 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerDied","Data":"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b"} Oct 03 14:38:55 crc kubenswrapper[4959]: I1003 14:38:55.761626 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerStarted","Data":"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149"} Oct 03 14:38:55 crc kubenswrapper[4959]: I1003 14:38:55.784626 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k6p82" podStartSLOduration=3.287240395 podStartE2EDuration="6.784605509s" podCreationTimestamp="2025-10-03 14:38:49 +0000 UTC" firstStartedPulling="2025-10-03 14:38:51.727710731 +0000 UTC m=+4100.931054138" lastFinishedPulling="2025-10-03 14:38:55.225075795 +0000 UTC m=+4104.428419252" observedRunningTime="2025-10-03 14:38:55.776857258 +0000 UTC m=+4104.980200675" watchObservedRunningTime="2025-10-03 14:38:55.784605509 +0000 UTC m=+4104.987948926" Oct 03 14:39:00 crc kubenswrapper[4959]: I1003 14:39:00.022360 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:00 crc kubenswrapper[4959]: I1003 14:39:00.022678 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:00 crc kubenswrapper[4959]: I1003 14:39:00.082944 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:00 crc kubenswrapper[4959]: I1003 14:39:00.840225 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:00 crc kubenswrapper[4959]: I1003 14:39:00.882412 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:39:02 crc kubenswrapper[4959]: I1003 14:39:02.811731 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k6p82" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="registry-server" containerID="cri-o://3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149" gracePeriod=2 Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.739291 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.822569 4959 generic.go:334] "Generic (PLEG): container finished" podID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerID="3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149" exitCode=0 Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.822607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerDied","Data":"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149"} Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.822633 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6p82" event={"ID":"df951cfb-e5e6-4aa5-b303-87407bf2daea","Type":"ContainerDied","Data":"a7ab56db71bb484ebe92d04ea000056108884785a7da4c1f1b8da5ab5222db02"} Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.822650 4959 scope.go:117] "RemoveContainer" containerID="3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.822811 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6p82" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.840597 4959 scope.go:117] "RemoveContainer" containerID="2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.857156 4959 scope.go:117] "RemoveContainer" containerID="db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.885624 4959 scope.go:117] "RemoveContainer" containerID="3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149" Oct 03 14:39:03 crc kubenswrapper[4959]: E1003 14:39:03.886138 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149\": container with ID starting with 3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149 not found: ID does not exist" containerID="3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.886179 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149"} err="failed to get container status \"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149\": rpc error: code = NotFound desc = could not find container \"3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149\": container with ID starting with 3c2e5e3bac77f2129395dfc3888aa2300bf34f508e449115c0baf6f0b3c55149 not found: ID does not exist" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.886218 4959 scope.go:117] "RemoveContainer" containerID="2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b" Oct 03 14:39:03 crc kubenswrapper[4959]: E1003 14:39:03.886635 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b\": container with ID starting with 2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b not found: ID does not exist" containerID="2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.886674 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b"} err="failed to get container status \"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b\": rpc error: code = NotFound desc = could not find container \"2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b\": container with ID starting with 2389d3efe71307de81bd97a5d43f71d7d6ff78d75ad1415925656f604d8ee40b not found: ID does not exist" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.886702 4959 scope.go:117] "RemoveContainer" containerID="db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d" Oct 03 14:39:03 crc kubenswrapper[4959]: E1003 14:39:03.887276 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d\": container with ID starting with db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d not found: ID does not exist" containerID="db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.887333 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d"} err="failed to get container status \"db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d\": rpc error: code = NotFound desc = could not find container \"db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d\": container with ID starting with db95841fbd2d90b999d921ed98d13f76265ac5cf5b35800b4b7afcf0285a559d not found: ID does not exist" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.915599 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities\") pod \"df951cfb-e5e6-4aa5-b303-87407bf2daea\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.915662 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tt4l\" (UniqueName: \"kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l\") pod \"df951cfb-e5e6-4aa5-b303-87407bf2daea\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.915735 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content\") pod \"df951cfb-e5e6-4aa5-b303-87407bf2daea\" (UID: \"df951cfb-e5e6-4aa5-b303-87407bf2daea\") " Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.916592 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities" (OuterVolumeSpecName: "utilities") pod "df951cfb-e5e6-4aa5-b303-87407bf2daea" (UID: "df951cfb-e5e6-4aa5-b303-87407bf2daea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:39:03 crc kubenswrapper[4959]: I1003 14:39:03.924303 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l" (OuterVolumeSpecName: "kube-api-access-8tt4l") pod "df951cfb-e5e6-4aa5-b303-87407bf2daea" (UID: "df951cfb-e5e6-4aa5-b303-87407bf2daea"). InnerVolumeSpecName "kube-api-access-8tt4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.012731 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df951cfb-e5e6-4aa5-b303-87407bf2daea" (UID: "df951cfb-e5e6-4aa5-b303-87407bf2daea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.016975 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.017029 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tt4l\" (UniqueName: \"kubernetes.io/projected/df951cfb-e5e6-4aa5-b303-87407bf2daea-kube-api-access-8tt4l\") on node \"crc\" DevicePath \"\"" Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.017044 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df951cfb-e5e6-4aa5-b303-87407bf2daea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.155420 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:39:04 crc kubenswrapper[4959]: I1003 14:39:04.161913 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k6p82"] Oct 03 14:39:05 crc kubenswrapper[4959]: I1003 14:39:05.694333 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" path="/var/lib/kubelet/pods/df951cfb-e5e6-4aa5-b303-87407bf2daea/volumes" Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.043940 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.043998 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.044055 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.044747 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.044820 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6" gracePeriod=600 Oct 03 14:39:06 crc kubenswrapper[4959]: E1003 14:39:06.200847 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6639a7dd_61b2_4eb7_bf37_2ccbe2487dde.slice/crio-conmon-9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6639a7dd_61b2_4eb7_bf37_2ccbe2487dde.slice/crio-9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6.scope\": RecentStats: unable to find data in memory cache]" Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.848235 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6" exitCode=0 Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.848330 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6"} Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.848609 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103"} Oct 03 14:39:06 crc kubenswrapper[4959]: I1003 14:39:06.848636 4959 scope.go:117] "RemoveContainer" containerID="96a8c3f6c9a85aae9693e44805ec35988e97b036a0a27f6da29fbc636d6069a5" Oct 03 14:41:06 crc kubenswrapper[4959]: I1003 14:41:06.044873 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:41:06 crc kubenswrapper[4959]: I1003 14:41:06.045403 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:41:36 crc kubenswrapper[4959]: I1003 14:41:36.044981 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:41:36 crc kubenswrapper[4959]: I1003 14:41:36.045638 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.044853 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.045421 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.045479 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.046186 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.046273 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" gracePeriod=600 Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.201923 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" exitCode=0 Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.201996 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103"} Oct 03 14:42:06 crc kubenswrapper[4959]: I1003 14:42:06.202284 4959 scope.go:117] "RemoveContainer" containerID="9f6a36ae07f21656f279545a9b2a1388d2232545f3431e3a167e23dda2ffc4e6" Oct 03 14:42:06 crc kubenswrapper[4959]: E1003 14:42:06.228329 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:42:07 crc kubenswrapper[4959]: I1003 14:42:07.215581 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:42:07 crc kubenswrapper[4959]: E1003 14:42:07.215939 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:42:21 crc kubenswrapper[4959]: I1003 14:42:21.691547 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:42:21 crc kubenswrapper[4959]: E1003 14:42:21.692558 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:42:32 crc kubenswrapper[4959]: I1003 14:42:32.685609 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:42:32 crc kubenswrapper[4959]: E1003 14:42:32.686573 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:42:45 crc kubenswrapper[4959]: I1003 14:42:45.685958 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:42:45 crc kubenswrapper[4959]: E1003 14:42:45.688868 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:42:57 crc kubenswrapper[4959]: I1003 14:42:57.685737 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:42:57 crc kubenswrapper[4959]: E1003 14:42:57.686586 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:43:09 crc kubenswrapper[4959]: I1003 14:43:09.685686 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:43:09 crc kubenswrapper[4959]: E1003 14:43:09.686627 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.309048 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:15 crc kubenswrapper[4959]: E1003 14:43:15.319464 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="extract-utilities" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.319573 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="extract-utilities" Oct 03 14:43:15 crc kubenswrapper[4959]: E1003 14:43:15.319642 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="extract-content" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.319707 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="extract-content" Oct 03 14:43:15 crc kubenswrapper[4959]: E1003 14:43:15.319806 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="registry-server" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.319877 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="registry-server" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.320095 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="df951cfb-e5e6-4aa5-b303-87407bf2daea" containerName="registry-server" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.327979 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.332095 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.376513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w945s\" (UniqueName: \"kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.376573 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.376616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.477422 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.477548 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w945s\" (UniqueName: \"kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.477581 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.477985 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.478135 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.497628 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w945s\" (UniqueName: \"kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s\") pod \"community-operators-vwkgx\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:15 crc kubenswrapper[4959]: I1003 14:43:15.665452 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:16 crc kubenswrapper[4959]: I1003 14:43:16.140279 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:16 crc kubenswrapper[4959]: I1003 14:43:16.725811 4959 generic.go:334] "Generic (PLEG): container finished" podID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerID="24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581" exitCode=0 Oct 03 14:43:16 crc kubenswrapper[4959]: I1003 14:43:16.725954 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerDied","Data":"24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581"} Oct 03 14:43:16 crc kubenswrapper[4959]: I1003 14:43:16.726206 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerStarted","Data":"f3260108ec2c4554aff31c8952e22006a54ed1e892a55e7b9664e9f9d5ad3674"} Oct 03 14:43:16 crc kubenswrapper[4959]: I1003 14:43:16.729108 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:43:18 crc kubenswrapper[4959]: I1003 14:43:18.740695 4959 generic.go:334] "Generic (PLEG): container finished" podID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerID="9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266" exitCode=0 Oct 03 14:43:18 crc kubenswrapper[4959]: I1003 14:43:18.741439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerDied","Data":"9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266"} Oct 03 14:43:19 crc kubenswrapper[4959]: I1003 14:43:19.753064 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerStarted","Data":"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9"} Oct 03 14:43:19 crc kubenswrapper[4959]: I1003 14:43:19.770802 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vwkgx" podStartSLOduration=2.317180566 podStartE2EDuration="4.770769852s" podCreationTimestamp="2025-10-03 14:43:15 +0000 UTC" firstStartedPulling="2025-10-03 14:43:16.728890096 +0000 UTC m=+4365.932233513" lastFinishedPulling="2025-10-03 14:43:19.182479382 +0000 UTC m=+4368.385822799" observedRunningTime="2025-10-03 14:43:19.770107326 +0000 UTC m=+4368.973450743" watchObservedRunningTime="2025-10-03 14:43:19.770769852 +0000 UTC m=+4368.974113269" Oct 03 14:43:24 crc kubenswrapper[4959]: I1003 14:43:24.685972 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:43:24 crc kubenswrapper[4959]: E1003 14:43:24.686764 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:43:25 crc kubenswrapper[4959]: I1003 14:43:25.666825 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:25 crc kubenswrapper[4959]: I1003 14:43:25.667963 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:25 crc kubenswrapper[4959]: I1003 14:43:25.819514 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:26 crc kubenswrapper[4959]: I1003 14:43:26.845465 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:26 crc kubenswrapper[4959]: I1003 14:43:26.898069 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:28 crc kubenswrapper[4959]: I1003 14:43:28.820501 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vwkgx" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="registry-server" containerID="cri-o://56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9" gracePeriod=2 Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.367236 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.472849 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w945s\" (UniqueName: \"kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s\") pod \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.472952 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities\") pod \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.472998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content\") pod \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\" (UID: \"2dd83fab-1b9b-47aa-b505-ec1215ac40d1\") " Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.474323 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities" (OuterVolumeSpecName: "utilities") pod "2dd83fab-1b9b-47aa-b505-ec1215ac40d1" (UID: "2dd83fab-1b9b-47aa-b505-ec1215ac40d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.486494 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s" (OuterVolumeSpecName: "kube-api-access-w945s") pod "2dd83fab-1b9b-47aa-b505-ec1215ac40d1" (UID: "2dd83fab-1b9b-47aa-b505-ec1215ac40d1"). InnerVolumeSpecName "kube-api-access-w945s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.529447 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dd83fab-1b9b-47aa-b505-ec1215ac40d1" (UID: "2dd83fab-1b9b-47aa-b505-ec1215ac40d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.574113 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.574149 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.574269 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w945s\" (UniqueName: \"kubernetes.io/projected/2dd83fab-1b9b-47aa-b505-ec1215ac40d1-kube-api-access-w945s\") on node \"crc\" DevicePath \"\"" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.830327 4959 generic.go:334] "Generic (PLEG): container finished" podID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerID="56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9" exitCode=0 Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.830391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerDied","Data":"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9"} Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.830420 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vwkgx" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.830696 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vwkgx" event={"ID":"2dd83fab-1b9b-47aa-b505-ec1215ac40d1","Type":"ContainerDied","Data":"f3260108ec2c4554aff31c8952e22006a54ed1e892a55e7b9664e9f9d5ad3674"} Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.830732 4959 scope.go:117] "RemoveContainer" containerID="56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.854577 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.865844 4959 scope.go:117] "RemoveContainer" containerID="9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.869949 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vwkgx"] Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.885329 4959 scope.go:117] "RemoveContainer" containerID="24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.912986 4959 scope.go:117] "RemoveContainer" containerID="56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9" Oct 03 14:43:29 crc kubenswrapper[4959]: E1003 14:43:29.913408 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9\": container with ID starting with 56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9 not found: ID does not exist" containerID="56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.913454 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9"} err="failed to get container status \"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9\": rpc error: code = NotFound desc = could not find container \"56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9\": container with ID starting with 56b187e1f3b7b78b95a959aaaa0b7da82482fa840cf042047f8b5a52b56ba7c9 not found: ID does not exist" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.913490 4959 scope.go:117] "RemoveContainer" containerID="9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266" Oct 03 14:43:29 crc kubenswrapper[4959]: E1003 14:43:29.913851 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266\": container with ID starting with 9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266 not found: ID does not exist" containerID="9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.913884 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266"} err="failed to get container status \"9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266\": rpc error: code = NotFound desc = could not find container \"9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266\": container with ID starting with 9e4f0ad679bea111124fe51760a40458dce2e61e66dfeca1147966ba1aadb266 not found: ID does not exist" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.913903 4959 scope.go:117] "RemoveContainer" containerID="24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581" Oct 03 14:43:29 crc kubenswrapper[4959]: E1003 14:43:29.914142 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581\": container with ID starting with 24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581 not found: ID does not exist" containerID="24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581" Oct 03 14:43:29 crc kubenswrapper[4959]: I1003 14:43:29.914170 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581"} err="failed to get container status \"24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581\": rpc error: code = NotFound desc = could not find container \"24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581\": container with ID starting with 24fada8422095ff4436ba5efd61ff545d396300accdc892ce24f4c56db940581 not found: ID does not exist" Oct 03 14:43:31 crc kubenswrapper[4959]: I1003 14:43:31.695066 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" path="/var/lib/kubelet/pods/2dd83fab-1b9b-47aa-b505-ec1215ac40d1/volumes" Oct 03 14:43:37 crc kubenswrapper[4959]: I1003 14:43:37.685177 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:43:37 crc kubenswrapper[4959]: E1003 14:43:37.685683 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:43:48 crc kubenswrapper[4959]: I1003 14:43:48.685328 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:43:48 crc kubenswrapper[4959]: E1003 14:43:48.686110 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:44:01 crc kubenswrapper[4959]: I1003 14:44:01.694582 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:44:01 crc kubenswrapper[4959]: E1003 14:44:01.695785 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:44:14 crc kubenswrapper[4959]: I1003 14:44:14.685031 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:44:14 crc kubenswrapper[4959]: E1003 14:44:14.685808 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:44:28 crc kubenswrapper[4959]: I1003 14:44:28.685827 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:44:28 crc kubenswrapper[4959]: E1003 14:44:28.686639 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:44:42 crc kubenswrapper[4959]: I1003 14:44:42.685617 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:44:42 crc kubenswrapper[4959]: E1003 14:44:42.686511 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:44:55 crc kubenswrapper[4959]: I1003 14:44:55.687073 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:44:55 crc kubenswrapper[4959]: E1003 14:44:55.688941 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.148751 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9"] Oct 03 14:45:00 crc kubenswrapper[4959]: E1003 14:45:00.149643 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="extract-utilities" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.149661 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="extract-utilities" Oct 03 14:45:00 crc kubenswrapper[4959]: E1003 14:45:00.149674 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="extract-content" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.149682 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="extract-content" Oct 03 14:45:00 crc kubenswrapper[4959]: E1003 14:45:00.149703 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="registry-server" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.149710 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="registry-server" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.149892 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd83fab-1b9b-47aa-b505-ec1215ac40d1" containerName="registry-server" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.150428 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.152576 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.153237 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.164970 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9"] Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.309146 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.309241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.309312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwlf6\" (UniqueName: \"kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.410836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwlf6\" (UniqueName: \"kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.410884 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.410925 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.413263 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.415888 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.439097 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwlf6\" (UniqueName: \"kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6\") pod \"collect-profiles-29325045-tmpv9\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.482797 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:00 crc kubenswrapper[4959]: I1003 14:45:00.940440 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9"] Oct 03 14:45:00 crc kubenswrapper[4959]: W1003 14:45:00.948382 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod015f6ed0_adc6_4144_8adf_bc69a828b80f.slice/crio-f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a WatchSource:0}: Error finding container f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a: Status 404 returned error can't find the container with id f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a Oct 03 14:45:01 crc kubenswrapper[4959]: I1003 14:45:01.596570 4959 generic.go:334] "Generic (PLEG): container finished" podID="015f6ed0-adc6-4144-8adf-bc69a828b80f" containerID="0ea57b1be2e4df0b5a9856387051928c05c85db2b6a2eb28bccc325e9f608fe2" exitCode=0 Oct 03 14:45:01 crc kubenswrapper[4959]: I1003 14:45:01.596646 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" event={"ID":"015f6ed0-adc6-4144-8adf-bc69a828b80f","Type":"ContainerDied","Data":"0ea57b1be2e4df0b5a9856387051928c05c85db2b6a2eb28bccc325e9f608fe2"} Oct 03 14:45:01 crc kubenswrapper[4959]: I1003 14:45:01.596986 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" event={"ID":"015f6ed0-adc6-4144-8adf-bc69a828b80f","Type":"ContainerStarted","Data":"f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a"} Oct 03 14:45:02 crc kubenswrapper[4959]: I1003 14:45:02.943694 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.047289 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume\") pod \"015f6ed0-adc6-4144-8adf-bc69a828b80f\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.047402 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume\") pod \"015f6ed0-adc6-4144-8adf-bc69a828b80f\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.047477 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwlf6\" (UniqueName: \"kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6\") pod \"015f6ed0-adc6-4144-8adf-bc69a828b80f\" (UID: \"015f6ed0-adc6-4144-8adf-bc69a828b80f\") " Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.048864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume" (OuterVolumeSpecName: "config-volume") pod "015f6ed0-adc6-4144-8adf-bc69a828b80f" (UID: "015f6ed0-adc6-4144-8adf-bc69a828b80f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.053541 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6" (OuterVolumeSpecName: "kube-api-access-jwlf6") pod "015f6ed0-adc6-4144-8adf-bc69a828b80f" (UID: "015f6ed0-adc6-4144-8adf-bc69a828b80f"). InnerVolumeSpecName "kube-api-access-jwlf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.053817 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "015f6ed0-adc6-4144-8adf-bc69a828b80f" (UID: "015f6ed0-adc6-4144-8adf-bc69a828b80f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.148813 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/015f6ed0-adc6-4144-8adf-bc69a828b80f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.148853 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/015f6ed0-adc6-4144-8adf-bc69a828b80f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.148867 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwlf6\" (UniqueName: \"kubernetes.io/projected/015f6ed0-adc6-4144-8adf-bc69a828b80f-kube-api-access-jwlf6\") on node \"crc\" DevicePath \"\"" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.610555 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" event={"ID":"015f6ed0-adc6-4144-8adf-bc69a828b80f","Type":"ContainerDied","Data":"f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a"} Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.610599 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f362d70626e8498c7bb050ebb619a3ca482cfcfbe080a537a88a723e1d26609a" Oct 03 14:45:03 crc kubenswrapper[4959]: I1003 14:45:03.610666 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9" Oct 03 14:45:04 crc kubenswrapper[4959]: I1003 14:45:04.033592 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5"] Oct 03 14:45:04 crc kubenswrapper[4959]: I1003 14:45:04.041746 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-b48l5"] Oct 03 14:45:05 crc kubenswrapper[4959]: I1003 14:45:05.696001 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1910e14-a1f9-4c81-a929-7336bd920153" path="/var/lib/kubelet/pods/a1910e14-a1f9-4c81-a929-7336bd920153/volumes" Oct 03 14:45:07 crc kubenswrapper[4959]: I1003 14:45:07.686095 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:45:07 crc kubenswrapper[4959]: E1003 14:45:07.686942 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:45:19 crc kubenswrapper[4959]: I1003 14:45:19.686369 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:45:19 crc kubenswrapper[4959]: E1003 14:45:19.686860 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:45:34 crc kubenswrapper[4959]: I1003 14:45:34.685207 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:45:34 crc kubenswrapper[4959]: E1003 14:45:34.685763 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:45:48 crc kubenswrapper[4959]: I1003 14:45:48.686111 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:45:48 crc kubenswrapper[4959]: E1003 14:45:48.687024 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:45:59 crc kubenswrapper[4959]: I1003 14:45:59.685925 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:45:59 crc kubenswrapper[4959]: E1003 14:45:59.687711 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:46:04 crc kubenswrapper[4959]: I1003 14:46:04.635719 4959 scope.go:117] "RemoveContainer" containerID="807e04c8d7efbb61d0e1d815bc66ec70315642f9ea107f603e0a535ab5061989" Oct 03 14:46:12 crc kubenswrapper[4959]: I1003 14:46:12.686311 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:46:12 crc kubenswrapper[4959]: E1003 14:46:12.687085 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.011430 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:16 crc kubenswrapper[4959]: E1003 14:46:16.011936 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015f6ed0-adc6-4144-8adf-bc69a828b80f" containerName="collect-profiles" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.011959 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="015f6ed0-adc6-4144-8adf-bc69a828b80f" containerName="collect-profiles" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.012329 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="015f6ed0-adc6-4144-8adf-bc69a828b80f" containerName="collect-profiles" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.014393 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.025740 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.142830 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.143059 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84v5q\" (UniqueName: \"kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.143111 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.244740 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.245325 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.245557 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.245831 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84v5q\" (UniqueName: \"kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.245837 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.272793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84v5q\" (UniqueName: \"kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q\") pod \"certified-operators-q5mbp\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.343591 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:16 crc kubenswrapper[4959]: I1003 14:46:16.818718 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:17 crc kubenswrapper[4959]: I1003 14:46:17.170291 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerID="ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09" exitCode=0 Oct 03 14:46:17 crc kubenswrapper[4959]: I1003 14:46:17.170378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerDied","Data":"ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09"} Oct 03 14:46:17 crc kubenswrapper[4959]: I1003 14:46:17.170595 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerStarted","Data":"71e14d3e2784e965cf1c6335c1547cb8170020be7272d3df83eccc980ff6bca7"} Oct 03 14:46:20 crc kubenswrapper[4959]: I1003 14:46:20.198350 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerID="bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0" exitCode=0 Oct 03 14:46:20 crc kubenswrapper[4959]: I1003 14:46:20.198397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerDied","Data":"bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0"} Oct 03 14:46:22 crc kubenswrapper[4959]: I1003 14:46:22.221214 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerStarted","Data":"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04"} Oct 03 14:46:22 crc kubenswrapper[4959]: I1003 14:46:22.249868 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q5mbp" podStartSLOduration=3.460288221 podStartE2EDuration="7.249847158s" podCreationTimestamp="2025-10-03 14:46:15 +0000 UTC" firstStartedPulling="2025-10-03 14:46:17.171785083 +0000 UTC m=+4546.375128500" lastFinishedPulling="2025-10-03 14:46:20.96134398 +0000 UTC m=+4550.164687437" observedRunningTime="2025-10-03 14:46:22.242924859 +0000 UTC m=+4551.446268286" watchObservedRunningTime="2025-10-03 14:46:22.249847158 +0000 UTC m=+4551.453190575" Oct 03 14:46:26 crc kubenswrapper[4959]: I1003 14:46:26.343763 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:26 crc kubenswrapper[4959]: I1003 14:46:26.344356 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:26 crc kubenswrapper[4959]: I1003 14:46:26.400305 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:27 crc kubenswrapper[4959]: I1003 14:46:27.304940 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:27 crc kubenswrapper[4959]: I1003 14:46:27.685895 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:46:27 crc kubenswrapper[4959]: E1003 14:46:27.686260 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:46:29 crc kubenswrapper[4959]: I1003 14:46:29.400061 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:29 crc kubenswrapper[4959]: I1003 14:46:29.400814 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q5mbp" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="registry-server" containerID="cri-o://767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04" gracePeriod=2 Oct 03 14:46:29 crc kubenswrapper[4959]: I1003 14:46:29.866227 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.043567 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content\") pod \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.043743 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84v5q\" (UniqueName: \"kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q\") pod \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.043807 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities\") pod \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\" (UID: \"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b\") " Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.045352 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities" (OuterVolumeSpecName: "utilities") pod "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" (UID: "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.048681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q" (OuterVolumeSpecName: "kube-api-access-84v5q") pod "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" (UID: "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b"). InnerVolumeSpecName "kube-api-access-84v5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.092051 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" (UID: "a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.145973 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84v5q\" (UniqueName: \"kubernetes.io/projected/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-kube-api-access-84v5q\") on node \"crc\" DevicePath \"\"" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.146051 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.146598 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.288623 4959 generic.go:334] "Generic (PLEG): container finished" podID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerID="767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04" exitCode=0 Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.288716 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerDied","Data":"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04"} Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.288748 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q5mbp" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.288794 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q5mbp" event={"ID":"a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b","Type":"ContainerDied","Data":"71e14d3e2784e965cf1c6335c1547cb8170020be7272d3df83eccc980ff6bca7"} Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.288857 4959 scope.go:117] "RemoveContainer" containerID="767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.326624 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.335074 4959 scope.go:117] "RemoveContainer" containerID="bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.340872 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q5mbp"] Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.373123 4959 scope.go:117] "RemoveContainer" containerID="ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.399600 4959 scope.go:117] "RemoveContainer" containerID="767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04" Oct 03 14:46:30 crc kubenswrapper[4959]: E1003 14:46:30.400129 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04\": container with ID starting with 767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04 not found: ID does not exist" containerID="767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.400171 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04"} err="failed to get container status \"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04\": rpc error: code = NotFound desc = could not find container \"767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04\": container with ID starting with 767af6392fe3b986fa2b28650f534c691c97f88faf252666f46756da9058af04 not found: ID does not exist" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.400258 4959 scope.go:117] "RemoveContainer" containerID="bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0" Oct 03 14:46:30 crc kubenswrapper[4959]: E1003 14:46:30.400662 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0\": container with ID starting with bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0 not found: ID does not exist" containerID="bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.400713 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0"} err="failed to get container status \"bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0\": rpc error: code = NotFound desc = could not find container \"bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0\": container with ID starting with bec2db7f2812b0a2f9c1d040ef8bcd4de7658911a387a59b885de4136f2da6c0 not found: ID does not exist" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.400750 4959 scope.go:117] "RemoveContainer" containerID="ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09" Oct 03 14:46:30 crc kubenswrapper[4959]: E1003 14:46:30.401096 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09\": container with ID starting with ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09 not found: ID does not exist" containerID="ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09" Oct 03 14:46:30 crc kubenswrapper[4959]: I1003 14:46:30.401123 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09"} err="failed to get container status \"ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09\": rpc error: code = NotFound desc = could not find container \"ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09\": container with ID starting with ca8fa43b2ffff940911af0c4a9585690ba112db1e70489befa65d09b34fc1c09 not found: ID does not exist" Oct 03 14:46:31 crc kubenswrapper[4959]: I1003 14:46:31.699325 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" path="/var/lib/kubelet/pods/a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b/volumes" Oct 03 14:46:39 crc kubenswrapper[4959]: I1003 14:46:39.685890 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:46:39 crc kubenswrapper[4959]: E1003 14:46:39.687762 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:46:50 crc kubenswrapper[4959]: I1003 14:46:50.685322 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:46:50 crc kubenswrapper[4959]: E1003 14:46:50.686012 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:47:02 crc kubenswrapper[4959]: I1003 14:47:02.685789 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:47:02 crc kubenswrapper[4959]: E1003 14:47:02.686679 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:47:13 crc kubenswrapper[4959]: I1003 14:47:13.685692 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:47:14 crc kubenswrapper[4959]: I1003 14:47:14.669033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4"} Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.486431 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-h5xbj"] Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.494142 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-h5xbj"] Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.607958 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-vvk92"] Oct 03 14:48:39 crc kubenswrapper[4959]: E1003 14:48:39.608530 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="extract-utilities" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.608572 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="extract-utilities" Oct 03 14:48:39 crc kubenswrapper[4959]: E1003 14:48:39.608594 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="registry-server" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.608611 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="registry-server" Oct 03 14:48:39 crc kubenswrapper[4959]: E1003 14:48:39.608656 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="extract-content" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.608676 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="extract-content" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.609045 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d44dba-1b4c-4779-8c3a-a77ef0b6fa3b" containerName="registry-server" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.610077 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.612327 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.612939 4959 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zbx7v" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.613947 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.614215 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.619497 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vvk92"] Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.693124 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19b5b78-580c-45b0-84a7-5fb0460a0114" path="/var/lib/kubelet/pods/c19b5b78-580c-45b0-84a7-5fb0460a0114/volumes" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.696506 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4qjx\" (UniqueName: \"kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.696644 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.696704 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.797801 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4qjx\" (UniqueName: \"kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.797961 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.798027 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.798568 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.799445 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.816493 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4qjx\" (UniqueName: \"kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx\") pod \"crc-storage-crc-vvk92\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:39 crc kubenswrapper[4959]: I1003 14:48:39.943244 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:40 crc kubenswrapper[4959]: I1003 14:48:40.366527 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vvk92"] Oct 03 14:48:40 crc kubenswrapper[4959]: I1003 14:48:40.383944 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:48:41 crc kubenswrapper[4959]: I1003 14:48:41.384586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vvk92" event={"ID":"34525299-f90e-4e94-8cc9-cb0d1162df9d","Type":"ContainerStarted","Data":"260d3d2b16636dd6f0ba6d7d7879202aebd1b1bcf5b558083c9496350c2d54c9"} Oct 03 14:48:41 crc kubenswrapper[4959]: I1003 14:48:41.384928 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vvk92" event={"ID":"34525299-f90e-4e94-8cc9-cb0d1162df9d","Type":"ContainerStarted","Data":"071193f9904842e74c363c27451af2764bab748cfad985751f8149f0c7d4b558"} Oct 03 14:48:41 crc kubenswrapper[4959]: I1003 14:48:41.402073 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-vvk92" podStartSLOduration=1.714292411 podStartE2EDuration="2.402045675s" podCreationTimestamp="2025-10-03 14:48:39 +0000 UTC" firstStartedPulling="2025-10-03 14:48:40.383720259 +0000 UTC m=+4689.587063676" lastFinishedPulling="2025-10-03 14:48:41.071473523 +0000 UTC m=+4690.274816940" observedRunningTime="2025-10-03 14:48:41.400662421 +0000 UTC m=+4690.604005848" watchObservedRunningTime="2025-10-03 14:48:41.402045675 +0000 UTC m=+4690.605389112" Oct 03 14:48:42 crc kubenswrapper[4959]: I1003 14:48:42.394722 4959 generic.go:334] "Generic (PLEG): container finished" podID="34525299-f90e-4e94-8cc9-cb0d1162df9d" containerID="260d3d2b16636dd6f0ba6d7d7879202aebd1b1bcf5b558083c9496350c2d54c9" exitCode=0 Oct 03 14:48:42 crc kubenswrapper[4959]: I1003 14:48:42.394784 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vvk92" event={"ID":"34525299-f90e-4e94-8cc9-cb0d1162df9d","Type":"ContainerDied","Data":"260d3d2b16636dd6f0ba6d7d7879202aebd1b1bcf5b558083c9496350c2d54c9"} Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.651484 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.760612 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt\") pod \"34525299-f90e-4e94-8cc9-cb0d1162df9d\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.760744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4qjx\" (UniqueName: \"kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx\") pod \"34525299-f90e-4e94-8cc9-cb0d1162df9d\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.760767 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "34525299-f90e-4e94-8cc9-cb0d1162df9d" (UID: "34525299-f90e-4e94-8cc9-cb0d1162df9d"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.760987 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage\") pod \"34525299-f90e-4e94-8cc9-cb0d1162df9d\" (UID: \"34525299-f90e-4e94-8cc9-cb0d1162df9d\") " Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.762451 4959 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/34525299-f90e-4e94-8cc9-cb0d1162df9d-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.773319 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx" (OuterVolumeSpecName: "kube-api-access-p4qjx") pod "34525299-f90e-4e94-8cc9-cb0d1162df9d" (UID: "34525299-f90e-4e94-8cc9-cb0d1162df9d"). InnerVolumeSpecName "kube-api-access-p4qjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.786115 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "34525299-f90e-4e94-8cc9-cb0d1162df9d" (UID: "34525299-f90e-4e94-8cc9-cb0d1162df9d"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.863766 4959 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/34525299-f90e-4e94-8cc9-cb0d1162df9d-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:43 crc kubenswrapper[4959]: I1003 14:48:43.863829 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4qjx\" (UniqueName: \"kubernetes.io/projected/34525299-f90e-4e94-8cc9-cb0d1162df9d-kube-api-access-p4qjx\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:44 crc kubenswrapper[4959]: I1003 14:48:44.413324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vvk92" event={"ID":"34525299-f90e-4e94-8cc9-cb0d1162df9d","Type":"ContainerDied","Data":"071193f9904842e74c363c27451af2764bab748cfad985751f8149f0c7d4b558"} Oct 03 14:48:44 crc kubenswrapper[4959]: I1003 14:48:44.413395 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="071193f9904842e74c363c27451af2764bab748cfad985751f8149f0c7d4b558" Oct 03 14:48:44 crc kubenswrapper[4959]: I1003 14:48:44.413419 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vvk92" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.524154 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-vvk92"] Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.535363 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-vvk92"] Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.654233 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-b6hh8"] Oct 03 14:48:45 crc kubenswrapper[4959]: E1003 14:48:45.654993 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34525299-f90e-4e94-8cc9-cb0d1162df9d" containerName="storage" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.655135 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="34525299-f90e-4e94-8cc9-cb0d1162df9d" containerName="storage" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.655580 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="34525299-f90e-4e94-8cc9-cb0d1162df9d" containerName="storage" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.656473 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.659250 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.660590 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-b6hh8"] Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.662675 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.662891 4959 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zbx7v" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.663098 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.705131 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34525299-f90e-4e94-8cc9-cb0d1162df9d" path="/var/lib/kubelet/pods/34525299-f90e-4e94-8cc9-cb0d1162df9d/volumes" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.796187 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.796444 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.796538 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxk9z\" (UniqueName: \"kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.898387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.898627 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.898695 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxk9z\" (UniqueName: \"kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.898728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.899314 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.917172 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxk9z\" (UniqueName: \"kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z\") pod \"crc-storage-crc-b6hh8\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:45 crc kubenswrapper[4959]: I1003 14:48:45.992749 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:46 crc kubenswrapper[4959]: I1003 14:48:46.289038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-b6hh8"] Oct 03 14:48:46 crc kubenswrapper[4959]: I1003 14:48:46.435780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-b6hh8" event={"ID":"5e687762-bce5-428c-994a-d23f1a470e59","Type":"ContainerStarted","Data":"57fcda2d5ea69032d1025c2928d3b12f679e6eb88023c16a279e01b1d63778d2"} Oct 03 14:48:47 crc kubenswrapper[4959]: I1003 14:48:47.446655 4959 generic.go:334] "Generic (PLEG): container finished" podID="5e687762-bce5-428c-994a-d23f1a470e59" containerID="8097f00420293c0cdab34038df7398c1720ca09c8a3dcb6b6a1533e353fe24c8" exitCode=0 Oct 03 14:48:47 crc kubenswrapper[4959]: I1003 14:48:47.446749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-b6hh8" event={"ID":"5e687762-bce5-428c-994a-d23f1a470e59","Type":"ContainerDied","Data":"8097f00420293c0cdab34038df7398c1720ca09c8a3dcb6b6a1533e353fe24c8"} Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.163462 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.269177 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt\") pod \"5e687762-bce5-428c-994a-d23f1a470e59\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.269258 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage\") pod \"5e687762-bce5-428c-994a-d23f1a470e59\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.269325 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxk9z\" (UniqueName: \"kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z\") pod \"5e687762-bce5-428c-994a-d23f1a470e59\" (UID: \"5e687762-bce5-428c-994a-d23f1a470e59\") " Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.269456 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "5e687762-bce5-428c-994a-d23f1a470e59" (UID: "5e687762-bce5-428c-994a-d23f1a470e59"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.269721 4959 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/5e687762-bce5-428c-994a-d23f1a470e59-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.277178 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z" (OuterVolumeSpecName: "kube-api-access-lxk9z") pod "5e687762-bce5-428c-994a-d23f1a470e59" (UID: "5e687762-bce5-428c-994a-d23f1a470e59"). InnerVolumeSpecName "kube-api-access-lxk9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.298533 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "5e687762-bce5-428c-994a-d23f1a470e59" (UID: "5e687762-bce5-428c-994a-d23f1a470e59"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.370875 4959 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/5e687762-bce5-428c-994a-d23f1a470e59-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.371180 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxk9z\" (UniqueName: \"kubernetes.io/projected/5e687762-bce5-428c-994a-d23f1a470e59-kube-api-access-lxk9z\") on node \"crc\" DevicePath \"\"" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.462676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-b6hh8" event={"ID":"5e687762-bce5-428c-994a-d23f1a470e59","Type":"ContainerDied","Data":"57fcda2d5ea69032d1025c2928d3b12f679e6eb88023c16a279e01b1d63778d2"} Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.462710 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57fcda2d5ea69032d1025c2928d3b12f679e6eb88023c16a279e01b1d63778d2" Oct 03 14:48:49 crc kubenswrapper[4959]: I1003 14:48:49.462759 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-b6hh8" Oct 03 14:49:04 crc kubenswrapper[4959]: I1003 14:49:04.747512 4959 scope.go:117] "RemoveContainer" containerID="2f78b669c90a683d26ba998dad174b1a18caeba81dcd7bda93930601ce904512" Oct 03 14:49:36 crc kubenswrapper[4959]: I1003 14:49:36.044671 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:49:36 crc kubenswrapper[4959]: I1003 14:49:36.045265 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:50:06 crc kubenswrapper[4959]: I1003 14:50:06.045161 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:50:06 crc kubenswrapper[4959]: I1003 14:50:06.045955 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.044466 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.045025 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.045086 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.046032 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.046135 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4" gracePeriod=600 Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.319822 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4" exitCode=0 Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.319866 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4"} Oct 03 14:50:36 crc kubenswrapper[4959]: I1003 14:50:36.319898 4959 scope.go:117] "RemoveContainer" containerID="8c68c379f0be64a959a67db591c0bbcc99133949d9d47543b0677cb0b4cd8103" Oct 03 14:50:37 crc kubenswrapper[4959]: I1003 14:50:37.329446 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f"} Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.079353 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:00 crc kubenswrapper[4959]: E1003 14:52:00.080726 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e687762-bce5-428c-994a-d23f1a470e59" containerName="storage" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.080760 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e687762-bce5-428c-994a-d23f1a470e59" containerName="storage" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.081143 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e687762-bce5-428c-994a-d23f1a470e59" containerName="storage" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.083455 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.087126 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.209775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.209825 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.209879 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgkb\" (UniqueName: \"kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.311417 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.311706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.311821 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgkb\" (UniqueName: \"kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.312707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.312707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.343278 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgkb\" (UniqueName: \"kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb\") pod \"redhat-marketplace-xhf28\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.409016 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:00 crc kubenswrapper[4959]: I1003 14:52:00.857017 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:01 crc kubenswrapper[4959]: I1003 14:52:01.038207 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerStarted","Data":"aece6746e1c92cb3ecd08b098daa629a6d58698ef209571e1be48bc8e6e59260"} Oct 03 14:52:02 crc kubenswrapper[4959]: I1003 14:52:02.046241 4959 generic.go:334] "Generic (PLEG): container finished" podID="90ffd741-ab01-48db-b7fe-11a89819844c" containerID="1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff" exitCode=0 Oct 03 14:52:02 crc kubenswrapper[4959]: I1003 14:52:02.046862 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerDied","Data":"1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff"} Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.057478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerStarted","Data":"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4"} Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.067806 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.069114 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.118858 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.155613 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.156817 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.159701 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-sffj8" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.159951 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.160064 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.160156 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.160278 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.168238 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.256818 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.256899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsw8s\" (UniqueName: \"kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.257000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh2rj\" (UniqueName: \"kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.257168 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.257272 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.257321 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358269 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358370 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsw8s\" (UniqueName: \"kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358407 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh2rj\" (UniqueName: \"kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358476 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.358522 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.359016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.359246 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.359333 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.359447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.379036 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsw8s\" (UniqueName: \"kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s\") pod \"redhat-operators-nvzzw\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.381245 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh2rj\" (UniqueName: \"kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj\") pod \"dnsmasq-dns-677fcb8757-k7mff\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.415674 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.416963 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.425703 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.437577 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.525632 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.560218 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.560311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.560331 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs76v\" (UniqueName: \"kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.664275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.664340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs76v\" (UniqueName: \"kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.664394 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.665267 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.665352 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.692784 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs76v\" (UniqueName: \"kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v\") pod \"dnsmasq-dns-98ddfc8f-62kvd\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.745494 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:03 crc kubenswrapper[4959]: I1003 14:52:03.970105 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:03 crc kubenswrapper[4959]: W1003 14:52:03.983050 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69f98cf7_4fa4_4a31_8634_7c865804461e.slice/crio-c49b5c7700f0d92e8cb6810d272e6ec4273e372e817a3af943757bff655f0ac6 WatchSource:0}: Error finding container c49b5c7700f0d92e8cb6810d272e6ec4273e372e817a3af943757bff655f0ac6: Status 404 returned error can't find the container with id c49b5c7700f0d92e8cb6810d272e6ec4273e372e817a3af943757bff655f0ac6 Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.018650 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.071980 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:04 crc kubenswrapper[4959]: W1003 14:52:04.075283 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9855eca_267e_4d64_a01a_664cb14ecbfb.slice/crio-4eb1b79cfeccf73c0ebfc065def44256c08fdfea26ccec91e665c2874918e606 WatchSource:0}: Error finding container 4eb1b79cfeccf73c0ebfc065def44256c08fdfea26ccec91e665c2874918e606: Status 404 returned error can't find the container with id 4eb1b79cfeccf73c0ebfc065def44256c08fdfea26ccec91e665c2874918e606 Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.077663 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" event={"ID":"7c81bc03-20dd-413e-a925-508ce945533b","Type":"ContainerStarted","Data":"e5927847ab9afa80b2428c08b0a81c523aba5e38c7ad89a832a13d2f418e769f"} Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.088343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerStarted","Data":"c49b5c7700f0d92e8cb6810d272e6ec4273e372e817a3af943757bff655f0ac6"} Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.091855 4959 generic.go:334] "Generic (PLEG): container finished" podID="90ffd741-ab01-48db-b7fe-11a89819844c" containerID="b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4" exitCode=0 Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.091906 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerDied","Data":"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4"} Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.293964 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.295676 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.299616 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-n8pl8" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.299855 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.299998 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.300058 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.300147 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.311040 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393546 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6zhs\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393595 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393630 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393697 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393726 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393825 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393850 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.393881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494827 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6zhs\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494881 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494914 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494945 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.494983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.495012 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.495036 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.495517 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.495571 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.496146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.496554 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.497955 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.497989 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/33cc606376ecf53ac77cc34db21598362dac9c5482bbc7da46869674823ec95a/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.500234 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.500283 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.501319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.516583 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6zhs\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.533965 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.565427 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.566834 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.569672 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.569923 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.570274 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.570430 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zn962" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.573138 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.600881 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.699950 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.700019 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.700472 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.700574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.700938 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.701044 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc66d\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.701261 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.701364 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.701465 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.741052 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.742331 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.747044 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.747256 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.747388 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.747522 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.750630 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.754104 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-57mzk" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.759626 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.768922 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814340 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-default\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814388 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814410 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814432 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-secrets\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814447 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-kolla-config\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814466 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814485 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814509 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814539 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814571 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814594 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814617 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckczq\" (UniqueName: \"kubernetes.io/projected/b654062a-f888-48b8-8b60-4fb64546dfe0-kube-api-access-ckczq\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814691 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814712 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.814757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc66d\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.815463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.820273 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.821042 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.822369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.825174 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.829724 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.834649 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.835099 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.836419 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ae9606a2b59bbdd3069f719969ed4004dad056963e449456f9876a6192c6b83f/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.847324 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc66d\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.857989 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.863724 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.868503 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2wrzw" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.868585 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.874086 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916437 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916486 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kolla-config\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916519 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckczq\" (UniqueName: \"kubernetes.io/projected/b654062a-f888-48b8-8b60-4fb64546dfe0-kube-api-access-ckczq\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916578 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-default\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916680 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-config-data\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m87vl\" (UniqueName: \"kubernetes.io/projected/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kube-api-access-m87vl\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-secrets\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-kolla-config\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916772 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.916802 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.922099 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-default\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.922681 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b654062a-f888-48b8-8b60-4fb64546dfe0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.923146 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.923617 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b654062a-f888-48b8-8b60-4fb64546dfe0-kolla-config\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.926453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.927466 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-secrets\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.927884 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b654062a-f888-48b8-8b60-4fb64546dfe0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.930770 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.930804 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e3ac43a12d2d8938513a88d1bf67ad2bfb1a1728ccefdb76e8adaeb681fc0cb0/globalmount\"" pod="openstack/openstack-galera-0" Oct 03 14:52:04 crc kubenswrapper[4959]: I1003 14:52:04.947590 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckczq\" (UniqueName: \"kubernetes.io/projected/b654062a-f888-48b8-8b60-4fb64546dfe0-kube-api-access-ckczq\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.007059 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " pod="openstack/rabbitmq-server-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.032689 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-config-data\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.032777 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m87vl\" (UniqueName: \"kubernetes.io/projected/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kube-api-access-m87vl\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.032917 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kolla-config\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.033912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kolla-config\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.034494 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57e3c964-7bc5-41a1-9c77-0f42cffb4628-config-data\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.047682 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74cdebb9-3b2d-4db4-a9ac-e8b70f011b04\") pod \"openstack-galera-0\" (UID: \"b654062a-f888-48b8-8b60-4fb64546dfe0\") " pod="openstack/openstack-galera-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.057957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m87vl\" (UniqueName: \"kubernetes.io/projected/57e3c964-7bc5-41a1-9c77-0f42cffb4628-kube-api-access-m87vl\") pod \"memcached-0\" (UID: \"57e3c964-7bc5-41a1-9c77-0f42cffb4628\") " pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.071595 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.109143 4959 generic.go:334] "Generic (PLEG): container finished" podID="7c81bc03-20dd-413e-a925-508ce945533b" containerID="52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4" exitCode=0 Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.109217 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" event={"ID":"7c81bc03-20dd-413e-a925-508ce945533b","Type":"ContainerDied","Data":"52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4"} Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.138384 4959 generic.go:334] "Generic (PLEG): container finished" podID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerID="2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67" exitCode=0 Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.138742 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" event={"ID":"f9855eca-267e-4d64-a01a-664cb14ecbfb","Type":"ContainerDied","Data":"2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67"} Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.138765 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" event={"ID":"f9855eca-267e-4d64-a01a-664cb14ecbfb","Type":"ContainerStarted","Data":"4eb1b79cfeccf73c0ebfc065def44256c08fdfea26ccec91e665c2874918e606"} Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.146108 4959 generic.go:334] "Generic (PLEG): container finished" podID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerID="238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6" exitCode=0 Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.147171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerDied","Data":"238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6"} Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.169682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerStarted","Data":"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c"} Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.190339 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.205613 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.233868 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhf28" podStartSLOduration=2.737375287 podStartE2EDuration="5.233845975s" podCreationTimestamp="2025-10-03 14:52:00 +0000 UTC" firstStartedPulling="2025-10-03 14:52:02.051586947 +0000 UTC m=+4891.254930374" lastFinishedPulling="2025-10-03 14:52:04.548057635 +0000 UTC m=+4893.751401062" observedRunningTime="2025-10-03 14:52:05.206572468 +0000 UTC m=+4894.409915895" watchObservedRunningTime="2025-10-03 14:52:05.233845975 +0000 UTC m=+4894.437189392" Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.334288 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:52:05 crc kubenswrapper[4959]: W1003 14:52:05.346001 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2435ee9c_4470_45c6_91f4_8045df435d77.slice/crio-fe87cad9bc09493385676ec05b229a7d911b431d97acd81a590ccfb4f3153172 WatchSource:0}: Error finding container fe87cad9bc09493385676ec05b229a7d911b431d97acd81a590ccfb4f3153172: Status 404 returned error can't find the container with id fe87cad9bc09493385676ec05b229a7d911b431d97acd81a590ccfb4f3153172 Oct 03 14:52:05 crc kubenswrapper[4959]: I1003 14:52:05.576244 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 14:52:05 crc kubenswrapper[4959]: W1003 14:52:05.593383 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb654062a_f888_48b8_8b60_4fb64546dfe0.slice/crio-037de3b421b234c4fd8142b5de990da0dd02963cf0e3086c05d1c47c642c67f9 WatchSource:0}: Error finding container 037de3b421b234c4fd8142b5de990da0dd02963cf0e3086c05d1c47c642c67f9: Status 404 returned error can't find the container with id 037de3b421b234c4fd8142b5de990da0dd02963cf0e3086c05d1c47c642c67f9 Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.628614 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 14:52:06 crc kubenswrapper[4959]: W1003 14:52:05.648663 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57e3c964_7bc5_41a1_9c77_0f42cffb4628.slice/crio-ada6f1f3db48844401473476b92456a14feedf485c68a78c5b4801d5fdf7930d WatchSource:0}: Error finding container ada6f1f3db48844401473476b92456a14feedf485c68a78c5b4801d5fdf7930d: Status 404 returned error can't find the container with id ada6f1f3db48844401473476b92456a14feedf485c68a78c5b4801d5fdf7930d Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.719700 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.721022 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.721157 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.725167 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.725233 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-m8558" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.725286 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.725239 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.797288 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:52:06 crc kubenswrapper[4959]: W1003 14:52:05.799828 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44ca5ab7_8cf3_43f4_a66c_f11444a793ec.slice/crio-8487e498182eb0ccdae194a81a475a9eab05c0e12fe1fcccc6365a9898701c77 WatchSource:0}: Error finding container 8487e498182eb0ccdae194a81a475a9eab05c0e12fe1fcccc6365a9898701c77: Status 404 returned error can't find the container with id 8487e498182eb0ccdae194a81a475a9eab05c0e12fe1fcccc6365a9898701c77 Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849271 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849306 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849326 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849348 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849372 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/de510575-5126-4c3a-9c24-15406b979fcf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frfqh\" (UniqueName: \"kubernetes.io/projected/de510575-5126-4c3a-9c24-15406b979fcf-kube-api-access-frfqh\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849483 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.849537 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.950664 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.950972 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.950989 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951066 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951095 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/de510575-5126-4c3a-9c24-15406b979fcf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951112 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frfqh\" (UniqueName: \"kubernetes.io/projected/de510575-5126-4c3a-9c24-15406b979fcf-kube-api-access-frfqh\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951130 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.951161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.952031 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.952155 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/de510575-5126-4c3a-9c24-15406b979fcf-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.952511 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.952809 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/de510575-5126-4c3a-9c24-15406b979fcf-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.954633 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.956000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.957032 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.957066 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a0b7613f7beb0210331bd1357930cf523ebb8269618d1cc7aeeb0e7e8845d8dd/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.958701 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de510575-5126-4c3a-9c24-15406b979fcf-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:05.973369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frfqh\" (UniqueName: \"kubernetes.io/projected/de510575-5126-4c3a-9c24-15406b979fcf-kube-api-access-frfqh\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.005773 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3457889-bf39-4e78-bd65-40c1a95a089f\") pod \"openstack-cell1-galera-0\" (UID: \"de510575-5126-4c3a-9c24-15406b979fcf\") " pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.055731 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.178104 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b654062a-f888-48b8-8b60-4fb64546dfe0","Type":"ContainerStarted","Data":"b2f8d989931a034dc037ef3f6bb2bea4f57a8cb59ec685aa0334776c0bbb38b0"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.178158 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b654062a-f888-48b8-8b60-4fb64546dfe0","Type":"ContainerStarted","Data":"037de3b421b234c4fd8142b5de990da0dd02963cf0e3086c05d1c47c642c67f9"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.180664 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerStarted","Data":"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.181996 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerStarted","Data":"8487e498182eb0ccdae194a81a475a9eab05c0e12fe1fcccc6365a9898701c77"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.182996 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerStarted","Data":"fe87cad9bc09493385676ec05b229a7d911b431d97acd81a590ccfb4f3153172"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.184663 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" event={"ID":"7c81bc03-20dd-413e-a925-508ce945533b","Type":"ContainerStarted","Data":"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.185368 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.187819 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" event={"ID":"f9855eca-267e-4d64-a01a-664cb14ecbfb","Type":"ContainerStarted","Data":"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.188252 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.189604 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"57e3c964-7bc5-41a1-9c77-0f42cffb4628","Type":"ContainerStarted","Data":"a7501770eaae3de1a324dd62562371cc984286814fabc99cbfb62c0cbaee45ae"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.189629 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"57e3c964-7bc5-41a1-9c77-0f42cffb4628","Type":"ContainerStarted","Data":"ada6f1f3db48844401473476b92456a14feedf485c68a78c5b4801d5fdf7930d"} Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.189717 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.225346 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.2253208349999998 podStartE2EDuration="2.225320835s" podCreationTimestamp="2025-10-03 14:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:06.22304675 +0000 UTC m=+4895.426390167" watchObservedRunningTime="2025-10-03 14:52:06.225320835 +0000 UTC m=+4895.428664252" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.258456 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" podStartSLOduration=3.258436956 podStartE2EDuration="3.258436956s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:06.254430608 +0000 UTC m=+4895.457774025" watchObservedRunningTime="2025-10-03 14:52:06.258436956 +0000 UTC m=+4895.461780373" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.278066 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" podStartSLOduration=3.278045026 podStartE2EDuration="3.278045026s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:06.269984638 +0000 UTC m=+4895.473328055" watchObservedRunningTime="2025-10-03 14:52:06.278045026 +0000 UTC m=+4895.481388443" Oct 03 14:52:06 crc kubenswrapper[4959]: I1003 14:52:06.525953 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 14:52:06 crc kubenswrapper[4959]: W1003 14:52:06.529828 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde510575_5126_4c3a_9c24_15406b979fcf.slice/crio-299d7d28f19db7a9714c26695cd84e5b600be17459516946569956c9bb3ae8c2 WatchSource:0}: Error finding container 299d7d28f19db7a9714c26695cd84e5b600be17459516946569956c9bb3ae8c2: Status 404 returned error can't find the container with id 299d7d28f19db7a9714c26695cd84e5b600be17459516946569956c9bb3ae8c2 Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.218915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"de510575-5126-4c3a-9c24-15406b979fcf","Type":"ContainerStarted","Data":"f9003f1175f19bd80c50051d82b917d41c6cf3aa9e2aea332814dfb73f5cb1c8"} Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.219251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"de510575-5126-4c3a-9c24-15406b979fcf","Type":"ContainerStarted","Data":"299d7d28f19db7a9714c26695cd84e5b600be17459516946569956c9bb3ae8c2"} Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.221338 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerStarted","Data":"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4"} Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.223781 4959 generic.go:334] "Generic (PLEG): container finished" podID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerID="d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832" exitCode=0 Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.223889 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerDied","Data":"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832"} Oct 03 14:52:07 crc kubenswrapper[4959]: I1003 14:52:07.225938 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerStarted","Data":"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12"} Oct 03 14:52:08 crc kubenswrapper[4959]: I1003 14:52:08.237300 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerStarted","Data":"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb"} Oct 03 14:52:08 crc kubenswrapper[4959]: I1003 14:52:08.270043 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nvzzw" podStartSLOduration=2.672102539 podStartE2EDuration="5.270020848s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="2025-10-03 14:52:05.148678801 +0000 UTC m=+4894.352022218" lastFinishedPulling="2025-10-03 14:52:07.74659711 +0000 UTC m=+4896.949940527" observedRunningTime="2025-10-03 14:52:08.263877827 +0000 UTC m=+4897.467221284" watchObservedRunningTime="2025-10-03 14:52:08.270020848 +0000 UTC m=+4897.473364275" Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.206870 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.252786 4959 generic.go:334] "Generic (PLEG): container finished" podID="de510575-5126-4c3a-9c24-15406b979fcf" containerID="f9003f1175f19bd80c50051d82b917d41c6cf3aa9e2aea332814dfb73f5cb1c8" exitCode=0 Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.252857 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"de510575-5126-4c3a-9c24-15406b979fcf","Type":"ContainerDied","Data":"f9003f1175f19bd80c50051d82b917d41c6cf3aa9e2aea332814dfb73f5cb1c8"} Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.254955 4959 generic.go:334] "Generic (PLEG): container finished" podID="b654062a-f888-48b8-8b60-4fb64546dfe0" containerID="b2f8d989931a034dc037ef3f6bb2bea4f57a8cb59ec685aa0334776c0bbb38b0" exitCode=0 Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.255010 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b654062a-f888-48b8-8b60-4fb64546dfe0","Type":"ContainerDied","Data":"b2f8d989931a034dc037ef3f6bb2bea4f57a8cb59ec685aa0334776c0bbb38b0"} Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.409519 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.409750 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:10 crc kubenswrapper[4959]: I1003 14:52:10.461061 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.275543 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"de510575-5126-4c3a-9c24-15406b979fcf","Type":"ContainerStarted","Data":"8c68c4f460c9579242f0c3decb9c7c19adc5fabe35b0f3673c69f1c574ce1df9"} Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.278581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b654062a-f888-48b8-8b60-4fb64546dfe0","Type":"ContainerStarted","Data":"1a8c3a92373133b56f43f0825eabb06eb9446110248039847338e2876c70ba5f"} Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.307574 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.307553313 podStartE2EDuration="7.307553313s" podCreationTimestamp="2025-10-03 14:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:11.301510345 +0000 UTC m=+4900.504853772" watchObservedRunningTime="2025-10-03 14:52:11.307553313 +0000 UTC m=+4900.510896750" Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.330046 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.348768 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.348735021 podStartE2EDuration="8.348735021s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:11.334926193 +0000 UTC m=+4900.538269610" watchObservedRunningTime="2025-10-03 14:52:11.348735021 +0000 UTC m=+4900.552078488" Oct 03 14:52:11 crc kubenswrapper[4959]: I1003 14:52:11.862891 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.294627 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhf28" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="registry-server" containerID="cri-o://e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c" gracePeriod=2 Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.439385 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.439479 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.492667 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.527409 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.747462 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:52:13 crc kubenswrapper[4959]: I1003 14:52:13.804038 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.255231 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.303758 4959 generic.go:334] "Generic (PLEG): container finished" podID="90ffd741-ab01-48db-b7fe-11a89819844c" containerID="e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c" exitCode=0 Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.304663 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhf28" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.305005 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerDied","Data":"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c"} Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.305031 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhf28" event={"ID":"90ffd741-ab01-48db-b7fe-11a89819844c","Type":"ContainerDied","Data":"aece6746e1c92cb3ecd08b098daa629a6d58698ef209571e1be48bc8e6e59260"} Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.305047 4959 scope.go:117] "RemoveContainer" containerID="e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.305220 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="dnsmasq-dns" containerID="cri-o://e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff" gracePeriod=10 Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.327117 4959 scope.go:117] "RemoveContainer" containerID="b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.344644 4959 scope.go:117] "RemoveContainer" containerID="1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.360231 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.380370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content\") pod \"90ffd741-ab01-48db-b7fe-11a89819844c\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.380429 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities\") pod \"90ffd741-ab01-48db-b7fe-11a89819844c\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.380531 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkgkb\" (UniqueName: \"kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb\") pod \"90ffd741-ab01-48db-b7fe-11a89819844c\" (UID: \"90ffd741-ab01-48db-b7fe-11a89819844c\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.381792 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities" (OuterVolumeSpecName: "utilities") pod "90ffd741-ab01-48db-b7fe-11a89819844c" (UID: "90ffd741-ab01-48db-b7fe-11a89819844c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.387273 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb" (OuterVolumeSpecName: "kube-api-access-bkgkb") pod "90ffd741-ab01-48db-b7fe-11a89819844c" (UID: "90ffd741-ab01-48db-b7fe-11a89819844c"). InnerVolumeSpecName "kube-api-access-bkgkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.398136 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90ffd741-ab01-48db-b7fe-11a89819844c" (UID: "90ffd741-ab01-48db-b7fe-11a89819844c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: E1003 14:52:14.404780 4959 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.245:47694->38.129.56.245:39745: read tcp 38.129.56.245:47694->38.129.56.245:39745: read: connection reset by peer Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.483227 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.483271 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ffd741-ab01-48db-b7fe-11a89819844c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.483291 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkgkb\" (UniqueName: \"kubernetes.io/projected/90ffd741-ab01-48db-b7fe-11a89819844c-kube-api-access-bkgkb\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.496068 4959 scope.go:117] "RemoveContainer" containerID="e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c" Oct 03 14:52:14 crc kubenswrapper[4959]: E1003 14:52:14.496604 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c\": container with ID starting with e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c not found: ID does not exist" containerID="e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.496651 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c"} err="failed to get container status \"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c\": rpc error: code = NotFound desc = could not find container \"e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c\": container with ID starting with e90e1e20cb78a81f1c5a68432ac27365cc933643612957521708df833768ec0c not found: ID does not exist" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.496681 4959 scope.go:117] "RemoveContainer" containerID="b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4" Oct 03 14:52:14 crc kubenswrapper[4959]: E1003 14:52:14.497086 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4\": container with ID starting with b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4 not found: ID does not exist" containerID="b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.497136 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4"} err="failed to get container status \"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4\": rpc error: code = NotFound desc = could not find container \"b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4\": container with ID starting with b8e384e6d6aa3f2011d3e3b2e5fc0c897c814adcdb49b19cbc88ead281019ab4 not found: ID does not exist" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.497163 4959 scope.go:117] "RemoveContainer" containerID="1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff" Oct 03 14:52:14 crc kubenswrapper[4959]: E1003 14:52:14.497719 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff\": container with ID starting with 1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff not found: ID does not exist" containerID="1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.497750 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff"} err="failed to get container status \"1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff\": rpc error: code = NotFound desc = could not find container \"1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff\": container with ID starting with 1aaf719f7ddfa51e7f569be14daa7870681c68f5092ee8a413e6045c381238ff not found: ID does not exist" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.641460 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.646768 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhf28"] Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.754620 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.890744 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh2rj\" (UniqueName: \"kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj\") pod \"f9855eca-267e-4d64-a01a-664cb14ecbfb\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.890851 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc\") pod \"f9855eca-267e-4d64-a01a-664cb14ecbfb\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.890948 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config\") pod \"f9855eca-267e-4d64-a01a-664cb14ecbfb\" (UID: \"f9855eca-267e-4d64-a01a-664cb14ecbfb\") " Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.897415 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj" (OuterVolumeSpecName: "kube-api-access-xh2rj") pod "f9855eca-267e-4d64-a01a-664cb14ecbfb" (UID: "f9855eca-267e-4d64-a01a-664cb14ecbfb"). InnerVolumeSpecName "kube-api-access-xh2rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.928969 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f9855eca-267e-4d64-a01a-664cb14ecbfb" (UID: "f9855eca-267e-4d64-a01a-664cb14ecbfb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.930156 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config" (OuterVolumeSpecName: "config") pod "f9855eca-267e-4d64-a01a-664cb14ecbfb" (UID: "f9855eca-267e-4d64-a01a-664cb14ecbfb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.992202 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.992579 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9855eca-267e-4d64-a01a-664cb14ecbfb-config\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:14 crc kubenswrapper[4959]: I1003 14:52:14.992594 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh2rj\" (UniqueName: \"kubernetes.io/projected/f9855eca-267e-4d64-a01a-664cb14ecbfb-kube-api-access-xh2rj\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.072916 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.073088 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.119001 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.321423 4959 generic.go:334] "Generic (PLEG): container finished" podID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerID="e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff" exitCode=0 Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.321518 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.321617 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" event={"ID":"f9855eca-267e-4d64-a01a-664cb14ecbfb","Type":"ContainerDied","Data":"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff"} Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.321658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677fcb8757-k7mff" event={"ID":"f9855eca-267e-4d64-a01a-664cb14ecbfb","Type":"ContainerDied","Data":"4eb1b79cfeccf73c0ebfc065def44256c08fdfea26ccec91e665c2874918e606"} Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.321683 4959 scope.go:117] "RemoveContainer" containerID="e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.343003 4959 scope.go:117] "RemoveContainer" containerID="2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.357860 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.365317 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-677fcb8757-k7mff"] Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.379586 4959 scope.go:117] "RemoveContainer" containerID="e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.380021 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 03 14:52:15 crc kubenswrapper[4959]: E1003 14:52:15.380183 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff\": container with ID starting with e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff not found: ID does not exist" containerID="e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.380261 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff"} err="failed to get container status \"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff\": rpc error: code = NotFound desc = could not find container \"e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff\": container with ID starting with e1050965e84da600ea8d5285ddd6b9ee59dcb1f9e29e9d3932e0dfc4178b76ff not found: ID does not exist" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.380297 4959 scope.go:117] "RemoveContainer" containerID="2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67" Oct 03 14:52:15 crc kubenswrapper[4959]: E1003 14:52:15.380709 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67\": container with ID starting with 2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67 not found: ID does not exist" containerID="2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.380739 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67"} err="failed to get container status \"2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67\": rpc error: code = NotFound desc = could not find container \"2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67\": container with ID starting with 2b5625630ddb3ced42717c8ab2b166fbbc542bbfcff2d55489e7d272ce8bac67 not found: ID does not exist" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.698088 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" path="/var/lib/kubelet/pods/90ffd741-ab01-48db-b7fe-11a89819844c/volumes" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.699076 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" path="/var/lib/kubelet/pods/f9855eca-267e-4d64-a01a-664cb14ecbfb/volumes" Oct 03 14:52:15 crc kubenswrapper[4959]: I1003 14:52:15.866720 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:16 crc kubenswrapper[4959]: I1003 14:52:16.056253 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:16 crc kubenswrapper[4959]: I1003 14:52:16.056307 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:16 crc kubenswrapper[4959]: I1003 14:52:16.332625 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nvzzw" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="registry-server" containerID="cri-o://6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb" gracePeriod=2 Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.294728 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.344092 4959 generic.go:334] "Generic (PLEG): container finished" podID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerID="6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb" exitCode=0 Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.344158 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nvzzw" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.344209 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerDied","Data":"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb"} Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.344239 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nvzzw" event={"ID":"69f98cf7-4fa4-4a31-8634-7c865804461e","Type":"ContainerDied","Data":"c49b5c7700f0d92e8cb6810d272e6ec4273e372e817a3af943757bff655f0ac6"} Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.344260 4959 scope.go:117] "RemoveContainer" containerID="6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.370644 4959 scope.go:117] "RemoveContainer" containerID="d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.408753 4959 scope.go:117] "RemoveContainer" containerID="238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.432738 4959 scope.go:117] "RemoveContainer" containerID="6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb" Oct 03 14:52:17 crc kubenswrapper[4959]: E1003 14:52:17.433721 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb\": container with ID starting with 6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb not found: ID does not exist" containerID="6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.433755 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb"} err="failed to get container status \"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb\": rpc error: code = NotFound desc = could not find container \"6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb\": container with ID starting with 6066c03b0998c27cb7a430fae7281c9e3dc2a169d38ad6100a1d57b1089b31cb not found: ID does not exist" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.433776 4959 scope.go:117] "RemoveContainer" containerID="d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832" Oct 03 14:52:17 crc kubenswrapper[4959]: E1003 14:52:17.434088 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832\": container with ID starting with d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832 not found: ID does not exist" containerID="d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.434127 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832"} err="failed to get container status \"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832\": rpc error: code = NotFound desc = could not find container \"d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832\": container with ID starting with d4f8af70a3eb5810af38148b6436f90a23b094deb8897c657244df1688ecb832 not found: ID does not exist" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.434159 4959 scope.go:117] "RemoveContainer" containerID="238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6" Oct 03 14:52:17 crc kubenswrapper[4959]: E1003 14:52:17.434797 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6\": container with ID starting with 238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6 not found: ID does not exist" containerID="238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.434822 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6"} err="failed to get container status \"238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6\": rpc error: code = NotFound desc = could not find container \"238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6\": container with ID starting with 238469daf3f7c25867aaee99558f9979b813eb8bd0cebc3965501f837f82f9a6 not found: ID does not exist" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.440693 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content\") pod \"69f98cf7-4fa4-4a31-8634-7c865804461e\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.440814 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsw8s\" (UniqueName: \"kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s\") pod \"69f98cf7-4fa4-4a31-8634-7c865804461e\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.440900 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities\") pod \"69f98cf7-4fa4-4a31-8634-7c865804461e\" (UID: \"69f98cf7-4fa4-4a31-8634-7c865804461e\") " Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.442091 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities" (OuterVolumeSpecName: "utilities") pod "69f98cf7-4fa4-4a31-8634-7c865804461e" (UID: "69f98cf7-4fa4-4a31-8634-7c865804461e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.445870 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s" (OuterVolumeSpecName: "kube-api-access-rsw8s") pod "69f98cf7-4fa4-4a31-8634-7c865804461e" (UID: "69f98cf7-4fa4-4a31-8634-7c865804461e"). InnerVolumeSpecName "kube-api-access-rsw8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.522745 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69f98cf7-4fa4-4a31-8634-7c865804461e" (UID: "69f98cf7-4fa4-4a31-8634-7c865804461e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.544608 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsw8s\" (UniqueName: \"kubernetes.io/projected/69f98cf7-4fa4-4a31-8634-7c865804461e-kube-api-access-rsw8s\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.544665 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.544677 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69f98cf7-4fa4-4a31-8634-7c865804461e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.696295 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:17 crc kubenswrapper[4959]: I1003 14:52:17.696330 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nvzzw"] Oct 03 14:52:18 crc kubenswrapper[4959]: I1003 14:52:18.144240 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:18 crc kubenswrapper[4959]: I1003 14:52:18.220925 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 03 14:52:19 crc kubenswrapper[4959]: I1003 14:52:19.698234 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" path="/var/lib/kubelet/pods/69f98cf7-4fa4-4a31-8634-7c865804461e/volumes" Oct 03 14:52:36 crc kubenswrapper[4959]: I1003 14:52:36.043898 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:52:36 crc kubenswrapper[4959]: I1003 14:52:36.044370 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:52:39 crc kubenswrapper[4959]: I1003 14:52:39.584413 4959 generic.go:334] "Generic (PLEG): container finished" podID="2435ee9c-4470-45c6-91f4-8045df435d77" containerID="023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4" exitCode=0 Oct 03 14:52:39 crc kubenswrapper[4959]: I1003 14:52:39.584659 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerDied","Data":"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4"} Oct 03 14:52:39 crc kubenswrapper[4959]: I1003 14:52:39.587048 4959 generic.go:334] "Generic (PLEG): container finished" podID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerID="db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12" exitCode=0 Oct 03 14:52:39 crc kubenswrapper[4959]: I1003 14:52:39.587070 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerDied","Data":"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12"} Oct 03 14:52:40 crc kubenswrapper[4959]: I1003 14:52:40.595628 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerStarted","Data":"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d"} Oct 03 14:52:40 crc kubenswrapper[4959]: I1003 14:52:40.596112 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:40 crc kubenswrapper[4959]: I1003 14:52:40.598039 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerStarted","Data":"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510"} Oct 03 14:52:40 crc kubenswrapper[4959]: I1003 14:52:40.598375 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 14:52:40 crc kubenswrapper[4959]: I1003 14:52:40.620462 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.620436092 podStartE2EDuration="37.620436092s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:40.618249868 +0000 UTC m=+4929.821593295" watchObservedRunningTime="2025-10-03 14:52:40.620436092 +0000 UTC m=+4929.823779519" Oct 03 14:52:54 crc kubenswrapper[4959]: I1003 14:52:54.754528 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:52:54 crc kubenswrapper[4959]: I1003 14:52:54.781594 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.781571655 podStartE2EDuration="51.781571655s" podCreationTimestamp="2025-10-03 14:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:52:40.643437664 +0000 UTC m=+4929.846781101" watchObservedRunningTime="2025-10-03 14:52:54.781571655 +0000 UTC m=+4943.984915082" Oct 03 14:52:55 crc kubenswrapper[4959]: I1003 14:52:55.194383 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425216 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425791 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="extract-utilities" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425805 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="extract-utilities" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425814 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="extract-utilities" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425820 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="extract-utilities" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425832 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="extract-content" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425838 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="extract-content" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425851 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="dnsmasq-dns" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425857 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="dnsmasq-dns" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425868 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425874 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425890 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425897 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425907 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="extract-content" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425912 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="extract-content" Oct 03 14:52:58 crc kubenswrapper[4959]: E1003 14:52:58.425922 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="init" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.425928 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="init" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.426075 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="90ffd741-ab01-48db-b7fe-11a89819844c" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.426093 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f98cf7-4fa4-4a31-8634-7c865804461e" containerName="registry-server" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.426104 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9855eca-267e-4d64-a01a-664cb14ecbfb" containerName="dnsmasq-dns" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.426870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.439811 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.553837 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.553893 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrdtq\" (UniqueName: \"kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.553930 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.655324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.655378 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrdtq\" (UniqueName: \"kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.655423 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.656236 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.656241 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.674255 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrdtq\" (UniqueName: \"kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq\") pod \"dnsmasq-dns-5b7946d7b9-964bk\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:58 crc kubenswrapper[4959]: I1003 14:52:58.746267 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.069621 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 14:52:59 crc kubenswrapper[4959]: W1003 14:52:59.074943 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aedf59f_7588_4d1c_ba17_fc60f3c960a6.slice/crio-3e44a15c397cef5227969d50ff1bba148b24337fc57fa533aff2b2a97cdfd823 WatchSource:0}: Error finding container 3e44a15c397cef5227969d50ff1bba148b24337fc57fa533aff2b2a97cdfd823: Status 404 returned error can't find the container with id 3e44a15c397cef5227969d50ff1bba148b24337fc57fa533aff2b2a97cdfd823 Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.177592 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.742389 4959 generic.go:334] "Generic (PLEG): container finished" podID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerID="0e6f6dc1a95bcbdfbb0ca3d70c425a54131ef3584e9efd961cf7e05a27eed5ac" exitCode=0 Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.742494 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" event={"ID":"1aedf59f-7588-4d1c-ba17-fc60f3c960a6","Type":"ContainerDied","Data":"0e6f6dc1a95bcbdfbb0ca3d70c425a54131ef3584e9efd961cf7e05a27eed5ac"} Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.743956 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" event={"ID":"1aedf59f-7588-4d1c-ba17-fc60f3c960a6","Type":"ContainerStarted","Data":"3e44a15c397cef5227969d50ff1bba148b24337fc57fa533aff2b2a97cdfd823"} Oct 03 14:52:59 crc kubenswrapper[4959]: I1003 14:52:59.879173 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:00 crc kubenswrapper[4959]: I1003 14:53:00.752822 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" event={"ID":"1aedf59f-7588-4d1c-ba17-fc60f3c960a6","Type":"ContainerStarted","Data":"d7c0b64eb42a181f5939991fa20a25043233055af0fd70d4eb99680609386d46"} Oct 03 14:53:00 crc kubenswrapper[4959]: I1003 14:53:00.753234 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:53:00 crc kubenswrapper[4959]: I1003 14:53:00.775359 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" podStartSLOduration=2.775342367 podStartE2EDuration="2.775342367s" podCreationTimestamp="2025-10-03 14:52:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:53:00.77016752 +0000 UTC m=+4949.973510967" watchObservedRunningTime="2025-10-03 14:53:00.775342367 +0000 UTC m=+4949.978685784" Oct 03 14:53:01 crc kubenswrapper[4959]: I1003 14:53:01.217402 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="rabbitmq" containerID="cri-o://7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510" gracePeriod=604798 Oct 03 14:53:01 crc kubenswrapper[4959]: I1003 14:53:01.791269 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="rabbitmq" containerID="cri-o://8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d" gracePeriod=604799 Oct 03 14:53:04 crc kubenswrapper[4959]: I1003 14:53:04.751812 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.241:5672: connect: connection refused" Oct 03 14:53:05 crc kubenswrapper[4959]: I1003 14:53:05.191346 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.242:5672: connect: connection refused" Oct 03 14:53:06 crc kubenswrapper[4959]: I1003 14:53:06.044923 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:53:06 crc kubenswrapper[4959]: I1003 14:53:06.045012 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.794570 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.811353 4959 generic.go:334] "Generic (PLEG): container finished" podID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerID="7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510" exitCode=0 Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.811402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerDied","Data":"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510"} Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.811434 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"44ca5ab7-8cf3-43f4-a66c-f11444a793ec","Type":"ContainerDied","Data":"8487e498182eb0ccdae194a81a475a9eab05c0e12fe1fcccc6365a9898701c77"} Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.811452 4959 scope.go:117] "RemoveContainer" containerID="7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.811575 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.848106 4959 scope.go:117] "RemoveContainer" containerID="db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.875763 4959 scope.go:117] "RemoveContainer" containerID="7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510" Oct 03 14:53:07 crc kubenswrapper[4959]: E1003 14:53:07.876217 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510\": container with ID starting with 7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510 not found: ID does not exist" containerID="7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.876248 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510"} err="failed to get container status \"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510\": rpc error: code = NotFound desc = could not find container \"7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510\": container with ID starting with 7008fc826e9683610b7b85a1765883ac2a4cca9fc4c87dbef0e5b417f3f58510 not found: ID does not exist" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.876273 4959 scope.go:117] "RemoveContainer" containerID="db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12" Oct 03 14:53:07 crc kubenswrapper[4959]: E1003 14:53:07.876529 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12\": container with ID starting with db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12 not found: ID does not exist" containerID="db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.876549 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12"} err="failed to get container status \"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12\": rpc error: code = NotFound desc = could not find container \"db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12\": container with ID starting with db243c7440cfeebe94569b353c094b95de61e7f5597aed1ed284e47f0986ef12 not found: ID does not exist" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908022 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908111 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908132 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908155 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908242 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908344 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908449 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908478 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908494 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc66d\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d\") pod \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\" (UID: \"44ca5ab7-8cf3-43f4-a66c-f11444a793ec\") " Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.908920 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.909093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.909348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.915713 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info" (OuterVolumeSpecName: "pod-info") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.927525 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.937549 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350" (OuterVolumeSpecName: "persistence") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.944017 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d" (OuterVolumeSpecName: "kube-api-access-gc66d") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "kube-api-access-gc66d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.956008 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf" (OuterVolumeSpecName: "server-conf") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:07 crc kubenswrapper[4959]: I1003 14:53:07.998350 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "44ca5ab7-8cf3-43f4-a66c-f11444a793ec" (UID: "44ca5ab7-8cf3-43f4-a66c-f11444a793ec"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010653 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010703 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc66d\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-kube-api-access-gc66d\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010722 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010742 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010759 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010774 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010791 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010839 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") on node \"crc\" " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.010854 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44ca5ab7-8cf3-43f4-a66c-f11444a793ec-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.027554 4959 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.027755 4959 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350") on node "crc" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.112495 4959 reconciler_common.go:293] "Volume detached for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.155538 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.160699 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.183919 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.184328 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.184354 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.184367 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="setup-container" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.184375 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="setup-container" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.184564 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.185510 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.187598 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.187968 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.188686 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zn962" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.189019 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.191124 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.197952 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.314754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.314930 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315121 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4csfp\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-kube-api-access-4csfp\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315186 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315280 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315341 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315370 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.315436 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416288 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416417 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416518 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4csfp\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-kube-api-access-4csfp\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416540 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416621 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.416647 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.417017 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.417977 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.418003 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.418367 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.418509 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.419140 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.419168 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ae9606a2b59bbdd3069f719969ed4004dad056963e449456f9876a6192c6b83f/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.420897 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.421421 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.422308 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.439329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4csfp\" (UniqueName: \"kubernetes.io/projected/dc4d11dc-89c5-49d3-a63b-44ef60b0489c-kube-api-access-4csfp\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.487480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e2cf41c-8011-41d9-83e4-aab3eabaa350\") pod \"rabbitmq-server-0\" (UID: \"dc4d11dc-89c5-49d3-a63b-44ef60b0489c\") " pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.509286 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.517891 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.518117 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.518371 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.518671 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.518731 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.519068 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.519133 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.519527 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.519613 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.519693 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.520047 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6zhs\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.520147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd\") pod \"2435ee9c-4470-45c6-91f4-8045df435d77\" (UID: \"2435ee9c-4470-45c6-91f4-8045df435d77\") " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.520563 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.520588 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.520602 4959 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.525338 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info" (OuterVolumeSpecName: "pod-info") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.525547 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs" (OuterVolumeSpecName: "kube-api-access-m6zhs") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "kube-api-access-m6zhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.533160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708" (OuterVolumeSpecName: "persistence") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "pvc-91586a76-c897-4a43-8f0f-38438e8c3708". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.536876 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf" (OuterVolumeSpecName: "server-conf") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.545881 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.622300 4959 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") on node \"crc\" " Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.622339 4959 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2435ee9c-4470-45c6-91f4-8045df435d77-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.622353 4959 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2435ee9c-4470-45c6-91f4-8045df435d77-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.622365 4959 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2435ee9c-4470-45c6-91f4-8045df435d77-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.622376 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6zhs\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-kube-api-access-m6zhs\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.630564 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2435ee9c-4470-45c6-91f4-8045df435d77" (UID: "2435ee9c-4470-45c6-91f4-8045df435d77"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.645534 4959 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.645693 4959 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-91586a76-c897-4a43-8f0f-38438e8c3708" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708") on node "crc" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.723869 4959 reconciler_common.go:293] "Volume detached for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.723916 4959 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2435ee9c-4470-45c6-91f4-8045df435d77-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.748867 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.801997 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.802314 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="dnsmasq-dns" containerID="cri-o://b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b" gracePeriod=10 Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.837857 4959 generic.go:334] "Generic (PLEG): container finished" podID="2435ee9c-4470-45c6-91f4-8045df435d77" containerID="8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d" exitCode=0 Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.837932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerDied","Data":"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d"} Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.837960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2435ee9c-4470-45c6-91f4-8045df435d77","Type":"ContainerDied","Data":"fe87cad9bc09493385676ec05b229a7d911b431d97acd81a590ccfb4f3153172"} Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.837986 4959 scope.go:117] "RemoveContainer" containerID="8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.838110 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.868498 4959 scope.go:117] "RemoveContainer" containerID="023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.932254 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.933354 4959 scope.go:117] "RemoveContainer" containerID="8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d" Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.938615 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d\": container with ID starting with 8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d not found: ID does not exist" containerID="8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.938669 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d"} err="failed to get container status \"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d\": rpc error: code = NotFound desc = could not find container \"8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d\": container with ID starting with 8dd83b838e6320d7f6516c0ae5419ce9bc9c3fd10c90748f93a0d4f16ae2584d not found: ID does not exist" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.938698 4959 scope.go:117] "RemoveContainer" containerID="023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.938807 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.944221 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4\": container with ID starting with 023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4 not found: ID does not exist" containerID="023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.944272 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4"} err="failed to get container status \"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4\": rpc error: code = NotFound desc = could not find container \"023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4\": container with ID starting with 023169acf7fee301e29ea22b5493259ee868ef0990b537d66d1ffbd3b9eae1f4 not found: ID does not exist" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.945846 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.946186 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.946219 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: E1003 14:53:08.946246 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="setup-container" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.946254 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="setup-container" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.946452 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" containerName="rabbitmq" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.959355 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.959448 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.962575 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.962738 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.963213 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.963649 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-n8pl8" Oct 03 14:53:08 crc kubenswrapper[4959]: I1003 14:53:08.964008 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036003 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036061 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036088 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036129 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036266 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036349 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4n2w\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-kube-api-access-s4n2w\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036456 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.036527 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.070633 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.137577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.137941 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.137976 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138089 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4n2w\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-kube-api-access-s4n2w\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138125 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138156 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.138214 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.139344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.139943 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.140344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.142469 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.149256 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.149683 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.152018 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.153164 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.153245 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/33cc606376ecf53ac77cc34db21598362dac9c5482bbc7da46869674823ec95a/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.163555 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4n2w\" (UniqueName: \"kubernetes.io/projected/6ac2cbd4-9c05-48d2-9c80-fad3bb420edc-kube-api-access-s4n2w\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.193840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-91586a76-c897-4a43-8f0f-38438e8c3708\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-91586a76-c897-4a43-8f0f-38438e8c3708\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.295556 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.304759 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.341125 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc\") pod \"7c81bc03-20dd-413e-a925-508ce945533b\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.341212 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs76v\" (UniqueName: \"kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v\") pod \"7c81bc03-20dd-413e-a925-508ce945533b\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.341262 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config\") pod \"7c81bc03-20dd-413e-a925-508ce945533b\" (UID: \"7c81bc03-20dd-413e-a925-508ce945533b\") " Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.347438 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v" (OuterVolumeSpecName: "kube-api-access-xs76v") pod "7c81bc03-20dd-413e-a925-508ce945533b" (UID: "7c81bc03-20dd-413e-a925-508ce945533b"). InnerVolumeSpecName "kube-api-access-xs76v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.375819 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c81bc03-20dd-413e-a925-508ce945533b" (UID: "7c81bc03-20dd-413e-a925-508ce945533b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.378423 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config" (OuterVolumeSpecName: "config") pod "7c81bc03-20dd-413e-a925-508ce945533b" (UID: "7c81bc03-20dd-413e-a925-508ce945533b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.443218 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.443610 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs76v\" (UniqueName: \"kubernetes.io/projected/7c81bc03-20dd-413e-a925-508ce945533b-kube-api-access-xs76v\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.443626 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c81bc03-20dd-413e-a925-508ce945533b-config\") on node \"crc\" DevicePath \"\"" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.695134 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2435ee9c-4470-45c6-91f4-8045df435d77" path="/var/lib/kubelet/pods/2435ee9c-4470-45c6-91f4-8045df435d77/volumes" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.695834 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44ca5ab7-8cf3-43f4-a66c-f11444a793ec" path="/var/lib/kubelet/pods/44ca5ab7-8cf3-43f4-a66c-f11444a793ec/volumes" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.772214 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 14:53:09 crc kubenswrapper[4959]: W1003 14:53:09.776656 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ac2cbd4_9c05_48d2_9c80_fad3bb420edc.slice/crio-f3f9661f8f08a0e5b26b1121ce048eea552de81de7f9d3ce854cbfecb28b8ad1 WatchSource:0}: Error finding container f3f9661f8f08a0e5b26b1121ce048eea552de81de7f9d3ce854cbfecb28b8ad1: Status 404 returned error can't find the container with id f3f9661f8f08a0e5b26b1121ce048eea552de81de7f9d3ce854cbfecb28b8ad1 Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.852167 4959 generic.go:334] "Generic (PLEG): container finished" podID="7c81bc03-20dd-413e-a925-508ce945533b" containerID="b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b" exitCode=0 Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.852243 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" event={"ID":"7c81bc03-20dd-413e-a925-508ce945533b","Type":"ContainerDied","Data":"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b"} Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.852275 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.852308 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-62kvd" event={"ID":"7c81bc03-20dd-413e-a925-508ce945533b","Type":"ContainerDied","Data":"e5927847ab9afa80b2428c08b0a81c523aba5e38c7ad89a832a13d2f418e769f"} Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.852335 4959 scope.go:117] "RemoveContainer" containerID="b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.854216 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc","Type":"ContainerStarted","Data":"f3f9661f8f08a0e5b26b1121ce048eea552de81de7f9d3ce854cbfecb28b8ad1"} Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.855878 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc4d11dc-89c5-49d3-a63b-44ef60b0489c","Type":"ContainerStarted","Data":"1fe45685ae9bf1a3fbbd5a102023cca5f39161627ce03afdef69fc1e84e6aab5"} Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.871474 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.872145 4959 scope.go:117] "RemoveContainer" containerID="52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4" Oct 03 14:53:09 crc kubenswrapper[4959]: I1003 14:53:09.876381 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-62kvd"] Oct 03 14:53:10 crc kubenswrapper[4959]: I1003 14:53:10.051500 4959 scope.go:117] "RemoveContainer" containerID="b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b" Oct 03 14:53:10 crc kubenswrapper[4959]: E1003 14:53:10.051935 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b\": container with ID starting with b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b not found: ID does not exist" containerID="b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b" Oct 03 14:53:10 crc kubenswrapper[4959]: I1003 14:53:10.051967 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b"} err="failed to get container status \"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b\": rpc error: code = NotFound desc = could not find container \"b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b\": container with ID starting with b12a3f4501ff8161a75d87053d64dcb630441ecb0fa94beba66062d841586a4b not found: ID does not exist" Oct 03 14:53:10 crc kubenswrapper[4959]: I1003 14:53:10.051987 4959 scope.go:117] "RemoveContainer" containerID="52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4" Oct 03 14:53:10 crc kubenswrapper[4959]: E1003 14:53:10.052754 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4\": container with ID starting with 52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4 not found: ID does not exist" containerID="52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4" Oct 03 14:53:10 crc kubenswrapper[4959]: I1003 14:53:10.052820 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4"} err="failed to get container status \"52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4\": rpc error: code = NotFound desc = could not find container \"52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4\": container with ID starting with 52252c2d946e46f1358ecc1ca7452f56d1f37aa937926bbc400b30e7255868e4 not found: ID does not exist" Oct 03 14:53:10 crc kubenswrapper[4959]: I1003 14:53:10.866368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc4d11dc-89c5-49d3-a63b-44ef60b0489c","Type":"ContainerStarted","Data":"26a18050a1d642fe23644e413636bbecb1f1a3cb8267622c7f7e158b1c923aee"} Oct 03 14:53:11 crc kubenswrapper[4959]: I1003 14:53:11.699441 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c81bc03-20dd-413e-a925-508ce945533b" path="/var/lib/kubelet/pods/7c81bc03-20dd-413e-a925-508ce945533b/volumes" Oct 03 14:53:11 crc kubenswrapper[4959]: I1003 14:53:11.878086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc","Type":"ContainerStarted","Data":"65e08429b055ae0848f7bcf928c451540f717b66a221856d754593937094c2a6"} Oct 03 14:53:36 crc kubenswrapper[4959]: I1003 14:53:36.045028 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:53:36 crc kubenswrapper[4959]: I1003 14:53:36.045691 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:53:36 crc kubenswrapper[4959]: I1003 14:53:36.045757 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 14:53:36 crc kubenswrapper[4959]: I1003 14:53:36.046763 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:53:36 crc kubenswrapper[4959]: I1003 14:53:36.046888 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" gracePeriod=600 Oct 03 14:53:36 crc kubenswrapper[4959]: E1003 14:53:36.181088 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:53:37 crc kubenswrapper[4959]: I1003 14:53:37.104027 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" exitCode=0 Oct 03 14:53:37 crc kubenswrapper[4959]: I1003 14:53:37.104081 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f"} Oct 03 14:53:37 crc kubenswrapper[4959]: I1003 14:53:37.104132 4959 scope.go:117] "RemoveContainer" containerID="632c4af9704f528e3a3e62fde688f4beb723642bfb74a5603174aecd3f0334a4" Oct 03 14:53:37 crc kubenswrapper[4959]: I1003 14:53:37.105070 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:53:37 crc kubenswrapper[4959]: E1003 14:53:37.105547 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:53:43 crc kubenswrapper[4959]: I1003 14:53:43.158280 4959 generic.go:334] "Generic (PLEG): container finished" podID="dc4d11dc-89c5-49d3-a63b-44ef60b0489c" containerID="26a18050a1d642fe23644e413636bbecb1f1a3cb8267622c7f7e158b1c923aee" exitCode=0 Oct 03 14:53:43 crc kubenswrapper[4959]: I1003 14:53:43.158397 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc4d11dc-89c5-49d3-a63b-44ef60b0489c","Type":"ContainerDied","Data":"26a18050a1d642fe23644e413636bbecb1f1a3cb8267622c7f7e158b1c923aee"} Oct 03 14:53:44 crc kubenswrapper[4959]: I1003 14:53:44.168949 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc4d11dc-89c5-49d3-a63b-44ef60b0489c","Type":"ContainerStarted","Data":"6fc44ec25f67ccb968b31a2d0387f178ccb424f7d1d407b615fb9a1596b10da6"} Oct 03 14:53:44 crc kubenswrapper[4959]: I1003 14:53:44.169596 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 14:53:44 crc kubenswrapper[4959]: I1003 14:53:44.170754 4959 generic.go:334] "Generic (PLEG): container finished" podID="6ac2cbd4-9c05-48d2-9c80-fad3bb420edc" containerID="65e08429b055ae0848f7bcf928c451540f717b66a221856d754593937094c2a6" exitCode=0 Oct 03 14:53:44 crc kubenswrapper[4959]: I1003 14:53:44.170805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc","Type":"ContainerDied","Data":"65e08429b055ae0848f7bcf928c451540f717b66a221856d754593937094c2a6"} Oct 03 14:53:44 crc kubenswrapper[4959]: I1003 14:53:44.208103 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.208078756 podStartE2EDuration="36.208078756s" podCreationTimestamp="2025-10-03 14:53:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:53:44.198731987 +0000 UTC m=+4993.402075434" watchObservedRunningTime="2025-10-03 14:53:44.208078756 +0000 UTC m=+4993.411422193" Oct 03 14:53:45 crc kubenswrapper[4959]: I1003 14:53:45.179593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ac2cbd4-9c05-48d2-9c80-fad3bb420edc","Type":"ContainerStarted","Data":"af51283a44ec4363f14a679b99e82702702adc78ac3d074180f461ec703c8b3a"} Oct 03 14:53:45 crc kubenswrapper[4959]: I1003 14:53:45.180318 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:53:45 crc kubenswrapper[4959]: I1003 14:53:45.203273 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.203257647 podStartE2EDuration="37.203257647s" podCreationTimestamp="2025-10-03 14:53:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:53:45.198897001 +0000 UTC m=+4994.402240428" watchObservedRunningTime="2025-10-03 14:53:45.203257647 +0000 UTC m=+4994.406601064" Oct 03 14:53:52 crc kubenswrapper[4959]: I1003 14:53:52.684838 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:53:52 crc kubenswrapper[4959]: E1003 14:53:52.685753 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:53:58 crc kubenswrapper[4959]: I1003 14:53:58.513579 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 14:53:59 crc kubenswrapper[4959]: I1003 14:53:59.308939 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 14:54:03 crc kubenswrapper[4959]: I1003 14:54:03.685850 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:54:03 crc kubenswrapper[4959]: E1003 14:54:03.686771 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.894948 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 03 14:54:07 crc kubenswrapper[4959]: E1003 14:54:07.895739 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="init" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.895764 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="init" Oct 03 14:54:07 crc kubenswrapper[4959]: E1003 14:54:07.895829 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="dnsmasq-dns" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.895842 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="dnsmasq-dns" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.896094 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c81bc03-20dd-413e-a925-508ce945533b" containerName="dnsmasq-dns" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.896889 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.899147 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vg28g" Oct 03 14:54:07 crc kubenswrapper[4959]: I1003 14:54:07.902680 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 03 14:54:08 crc kubenswrapper[4959]: I1003 14:54:08.068775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ll7s\" (UniqueName: \"kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s\") pod \"mariadb-client-1-default\" (UID: \"2177fd84-0946-4734-a233-ad8f54d46238\") " pod="openstack/mariadb-client-1-default" Oct 03 14:54:08 crc kubenswrapper[4959]: I1003 14:54:08.170295 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ll7s\" (UniqueName: \"kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s\") pod \"mariadb-client-1-default\" (UID: \"2177fd84-0946-4734-a233-ad8f54d46238\") " pod="openstack/mariadb-client-1-default" Oct 03 14:54:08 crc kubenswrapper[4959]: I1003 14:54:08.190134 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ll7s\" (UniqueName: \"kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s\") pod \"mariadb-client-1-default\" (UID: \"2177fd84-0946-4734-a233-ad8f54d46238\") " pod="openstack/mariadb-client-1-default" Oct 03 14:54:08 crc kubenswrapper[4959]: I1003 14:54:08.231245 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 03 14:54:08 crc kubenswrapper[4959]: I1003 14:54:08.752026 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 03 14:54:08 crc kubenswrapper[4959]: W1003 14:54:08.755009 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2177fd84_0946_4734_a233_ad8f54d46238.slice/crio-67e11f3573e7f8dc73cea605ad692ef31e848d132ff7e7c44c1e10bec9e9ea9c WatchSource:0}: Error finding container 67e11f3573e7f8dc73cea605ad692ef31e848d132ff7e7c44c1e10bec9e9ea9c: Status 404 returned error can't find the container with id 67e11f3573e7f8dc73cea605ad692ef31e848d132ff7e7c44c1e10bec9e9ea9c Oct 03 14:54:09 crc kubenswrapper[4959]: I1003 14:54:09.395716 4959 generic.go:334] "Generic (PLEG): container finished" podID="2177fd84-0946-4734-a233-ad8f54d46238" containerID="72959b2d8bc8e0fb5b2ae1654f4cc40d682e221747c87180c11a01f00c8aab1f" exitCode=0 Oct 03 14:54:09 crc kubenswrapper[4959]: I1003 14:54:09.395803 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"2177fd84-0946-4734-a233-ad8f54d46238","Type":"ContainerDied","Data":"72959b2d8bc8e0fb5b2ae1654f4cc40d682e221747c87180c11a01f00c8aab1f"} Oct 03 14:54:09 crc kubenswrapper[4959]: I1003 14:54:09.396068 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"2177fd84-0946-4734-a233-ad8f54d46238","Type":"ContainerStarted","Data":"67e11f3573e7f8dc73cea605ad692ef31e848d132ff7e7c44c1e10bec9e9ea9c"} Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.775475 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.839449 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_2177fd84-0946-4734-a233-ad8f54d46238/mariadb-client-1-default/0.log" Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.864376 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.872077 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.928270 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ll7s\" (UniqueName: \"kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s\") pod \"2177fd84-0946-4734-a233-ad8f54d46238\" (UID: \"2177fd84-0946-4734-a233-ad8f54d46238\") " Oct 03 14:54:10 crc kubenswrapper[4959]: I1003 14:54:10.933441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s" (OuterVolumeSpecName: "kube-api-access-7ll7s") pod "2177fd84-0946-4734-a233-ad8f54d46238" (UID: "2177fd84-0946-4734-a233-ad8f54d46238"). InnerVolumeSpecName "kube-api-access-7ll7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.030969 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ll7s\" (UniqueName: \"kubernetes.io/projected/2177fd84-0946-4734-a233-ad8f54d46238-kube-api-access-7ll7s\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.298932 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 03 14:54:11 crc kubenswrapper[4959]: E1003 14:54:11.299437 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2177fd84-0946-4734-a233-ad8f54d46238" containerName="mariadb-client-1-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.299462 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2177fd84-0946-4734-a233-ad8f54d46238" containerName="mariadb-client-1-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.299650 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2177fd84-0946-4734-a233-ad8f54d46238" containerName="mariadb-client-1-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.300302 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.307295 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.334364 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws2g5\" (UniqueName: \"kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5\") pod \"mariadb-client-2-default\" (UID: \"f6edeb3f-ab59-40fb-8d3e-d72839714b66\") " pod="openstack/mariadb-client-2-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.411708 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e11f3573e7f8dc73cea605ad692ef31e848d132ff7e7c44c1e10bec9e9ea9c" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.411787 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.435785 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws2g5\" (UniqueName: \"kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5\") pod \"mariadb-client-2-default\" (UID: \"f6edeb3f-ab59-40fb-8d3e-d72839714b66\") " pod="openstack/mariadb-client-2-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.452085 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws2g5\" (UniqueName: \"kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5\") pod \"mariadb-client-2-default\" (UID: \"f6edeb3f-ab59-40fb-8d3e-d72839714b66\") " pod="openstack/mariadb-client-2-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.627746 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.703687 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2177fd84-0946-4734-a233-ad8f54d46238" path="/var/lib/kubelet/pods/2177fd84-0946-4734-a233-ad8f54d46238/volumes" Oct 03 14:54:11 crc kubenswrapper[4959]: I1003 14:54:11.981898 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 03 14:54:11 crc kubenswrapper[4959]: W1003 14:54:11.988215 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6edeb3f_ab59_40fb_8d3e_d72839714b66.slice/crio-c38181df6dbff3710002cc0d86c0b3dca837f2916ad08ef6602e149df8c8f184 WatchSource:0}: Error finding container c38181df6dbff3710002cc0d86c0b3dca837f2916ad08ef6602e149df8c8f184: Status 404 returned error can't find the container with id c38181df6dbff3710002cc0d86c0b3dca837f2916ad08ef6602e149df8c8f184 Oct 03 14:54:12 crc kubenswrapper[4959]: I1003 14:54:12.420471 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"f6edeb3f-ab59-40fb-8d3e-d72839714b66","Type":"ContainerStarted","Data":"b4dcf4da2beface3decaa5295dd894e223f459d934ac2db231f3a284089363bd"} Oct 03 14:54:12 crc kubenswrapper[4959]: I1003 14:54:12.420514 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"f6edeb3f-ab59-40fb-8d3e-d72839714b66","Type":"ContainerStarted","Data":"c38181df6dbff3710002cc0d86c0b3dca837f2916ad08ef6602e149df8c8f184"} Oct 03 14:54:12 crc kubenswrapper[4959]: I1003 14:54:12.445070 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.445037739 podStartE2EDuration="1.445037739s" podCreationTimestamp="2025-10-03 14:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:54:12.436645444 +0000 UTC m=+5021.639988891" watchObservedRunningTime="2025-10-03 14:54:12.445037739 +0000 UTC m=+5021.648381196" Oct 03 14:54:12 crc kubenswrapper[4959]: E1003 14:54:12.922323 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:34866->38.129.56.245:39745: write tcp 38.129.56.245:34866->38.129.56.245:39745: write: broken pipe Oct 03 14:54:13 crc kubenswrapper[4959]: I1003 14:54:13.430493 4959 generic.go:334] "Generic (PLEG): container finished" podID="f6edeb3f-ab59-40fb-8d3e-d72839714b66" containerID="b4dcf4da2beface3decaa5295dd894e223f459d934ac2db231f3a284089363bd" exitCode=0 Oct 03 14:54:13 crc kubenswrapper[4959]: I1003 14:54:13.430544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"f6edeb3f-ab59-40fb-8d3e-d72839714b66","Type":"ContainerDied","Data":"b4dcf4da2beface3decaa5295dd894e223f459d934ac2db231f3a284089363bd"} Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.781538 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.814034 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.818367 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.908440 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws2g5\" (UniqueName: \"kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5\") pod \"f6edeb3f-ab59-40fb-8d3e-d72839714b66\" (UID: \"f6edeb3f-ab59-40fb-8d3e-d72839714b66\") " Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.916471 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5" (OuterVolumeSpecName: "kube-api-access-ws2g5") pod "f6edeb3f-ab59-40fb-8d3e-d72839714b66" (UID: "f6edeb3f-ab59-40fb-8d3e-d72839714b66"). InnerVolumeSpecName "kube-api-access-ws2g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.963860 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-3-default"] Oct 03 14:54:14 crc kubenswrapper[4959]: E1003 14:54:14.964382 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6edeb3f-ab59-40fb-8d3e-d72839714b66" containerName="mariadb-client-2-default" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.964407 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6edeb3f-ab59-40fb-8d3e-d72839714b66" containerName="mariadb-client-2-default" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.964726 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6edeb3f-ab59-40fb-8d3e-d72839714b66" containerName="mariadb-client-2-default" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.965993 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 03 14:54:14 crc kubenswrapper[4959]: I1003 14:54:14.969989 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.010361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtwv\" (UniqueName: \"kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv\") pod \"mariadb-client-3-default\" (UID: \"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db\") " pod="openstack/mariadb-client-3-default" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.010582 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws2g5\" (UniqueName: \"kubernetes.io/projected/f6edeb3f-ab59-40fb-8d3e-d72839714b66-kube-api-access-ws2g5\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.111883 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtwv\" (UniqueName: \"kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv\") pod \"mariadb-client-3-default\" (UID: \"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db\") " pod="openstack/mariadb-client-3-default" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.128512 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtwv\" (UniqueName: \"kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv\") pod \"mariadb-client-3-default\" (UID: \"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db\") " pod="openstack/mariadb-client-3-default" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.305216 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.466703 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38181df6dbff3710002cc0d86c0b3dca837f2916ad08ef6602e149df8c8f184" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.466743 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.697600 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6edeb3f-ab59-40fb-8d3e-d72839714b66" path="/var/lib/kubelet/pods/f6edeb3f-ab59-40fb-8d3e-d72839714b66/volumes" Oct 03 14:54:15 crc kubenswrapper[4959]: I1003 14:54:15.862334 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 03 14:54:15 crc kubenswrapper[4959]: W1003 14:54:15.866259 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8744a2d_1f5f_4cc8_b917_61d7ad86e8db.slice/crio-df247c751c36adc877c19fa63a22c55b1993c21329425fcc063e72ce85b47a9c WatchSource:0}: Error finding container df247c751c36adc877c19fa63a22c55b1993c21329425fcc063e72ce85b47a9c: Status 404 returned error can't find the container with id df247c751c36adc877c19fa63a22c55b1993c21329425fcc063e72ce85b47a9c Oct 03 14:54:16 crc kubenswrapper[4959]: I1003 14:54:16.479433 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db","Type":"ContainerStarted","Data":"6e1623d214c36de6c2178ff445457c448cc21feacc3f66e6e1a893c43e1db9d6"} Oct 03 14:54:16 crc kubenswrapper[4959]: I1003 14:54:16.479833 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db","Type":"ContainerStarted","Data":"df247c751c36adc877c19fa63a22c55b1993c21329425fcc063e72ce85b47a9c"} Oct 03 14:54:16 crc kubenswrapper[4959]: I1003 14:54:16.503261 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-3-default" podStartSLOduration=2.50322688 podStartE2EDuration="2.50322688s" podCreationTimestamp="2025-10-03 14:54:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:54:16.496145727 +0000 UTC m=+5025.699489194" watchObservedRunningTime="2025-10-03 14:54:16.50322688 +0000 UTC m=+5025.706570337" Oct 03 14:54:18 crc kubenswrapper[4959]: I1003 14:54:18.502477 4959 generic.go:334] "Generic (PLEG): container finished" podID="e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" containerID="6e1623d214c36de6c2178ff445457c448cc21feacc3f66e6e1a893c43e1db9d6" exitCode=0 Oct 03 14:54:18 crc kubenswrapper[4959]: I1003 14:54:18.502552 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-3-default" event={"ID":"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db","Type":"ContainerDied","Data":"6e1623d214c36de6c2178ff445457c448cc21feacc3f66e6e1a893c43e1db9d6"} Oct 03 14:54:18 crc kubenswrapper[4959]: I1003 14:54:18.686022 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:54:18 crc kubenswrapper[4959]: E1003 14:54:18.686428 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:54:19 crc kubenswrapper[4959]: I1003 14:54:19.965479 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.009932 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.019534 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-3-default"] Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.089694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbtwv\" (UniqueName: \"kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv\") pod \"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db\" (UID: \"e8744a2d-1f5f-4cc8-b917-61d7ad86e8db\") " Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.096231 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv" (OuterVolumeSpecName: "kube-api-access-qbtwv") pod "e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" (UID: "e8744a2d-1f5f-4cc8-b917-61d7ad86e8db"). InnerVolumeSpecName "kube-api-access-qbtwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.196161 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbtwv\" (UniqueName: \"kubernetes.io/projected/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db-kube-api-access-qbtwv\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.478846 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 03 14:54:20 crc kubenswrapper[4959]: E1003 14:54:20.479244 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" containerName="mariadb-client-3-default" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.479268 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" containerName="mariadb-client-3-default" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.479521 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" containerName="mariadb-client-3-default" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.480208 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.493719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.500618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjmm5\" (UniqueName: \"kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5\") pod \"mariadb-client-1\" (UID: \"ad2223ee-aa7f-434c-9f0d-8da7c8071f48\") " pod="openstack/mariadb-client-1" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.523303 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df247c751c36adc877c19fa63a22c55b1993c21329425fcc063e72ce85b47a9c" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.523398 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-3-default" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.601911 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjmm5\" (UniqueName: \"kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5\") pod \"mariadb-client-1\" (UID: \"ad2223ee-aa7f-434c-9f0d-8da7c8071f48\") " pod="openstack/mariadb-client-1" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.621398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjmm5\" (UniqueName: \"kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5\") pod \"mariadb-client-1\" (UID: \"ad2223ee-aa7f-434c-9f0d-8da7c8071f48\") " pod="openstack/mariadb-client-1" Oct 03 14:54:20 crc kubenswrapper[4959]: I1003 14:54:20.823459 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.356370 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.539132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ad2223ee-aa7f-434c-9f0d-8da7c8071f48","Type":"ContainerStarted","Data":"aaab30786b377eecb4418fc42d5429a04cbd04b9ff4646fcce80433647bde683"} Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.539538 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ad2223ee-aa7f-434c-9f0d-8da7c8071f48","Type":"ContainerStarted","Data":"fe0f6aad187520c6018b0dd4798e7988bc3128e01e5e2b2e86ded2b8b32e2a02"} Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.552241 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-1" podStartSLOduration=1.552224115 podStartE2EDuration="1.552224115s" podCreationTimestamp="2025-10-03 14:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:54:21.551780123 +0000 UTC m=+5030.755123540" watchObservedRunningTime="2025-10-03 14:54:21.552224115 +0000 UTC m=+5030.755567522" Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.605253 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_ad2223ee-aa7f-434c-9f0d-8da7c8071f48/mariadb-client-1/0.log" Oct 03 14:54:21 crc kubenswrapper[4959]: I1003 14:54:21.699158 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8744a2d-1f5f-4cc8-b917-61d7ad86e8db" path="/var/lib/kubelet/pods/e8744a2d-1f5f-4cc8-b917-61d7ad86e8db/volumes" Oct 03 14:54:22 crc kubenswrapper[4959]: I1003 14:54:22.551501 4959 generic.go:334] "Generic (PLEG): container finished" podID="ad2223ee-aa7f-434c-9f0d-8da7c8071f48" containerID="aaab30786b377eecb4418fc42d5429a04cbd04b9ff4646fcce80433647bde683" exitCode=0 Oct 03 14:54:22 crc kubenswrapper[4959]: I1003 14:54:22.551555 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"ad2223ee-aa7f-434c-9f0d-8da7c8071f48","Type":"ContainerDied","Data":"aaab30786b377eecb4418fc42d5429a04cbd04b9ff4646fcce80433647bde683"} Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.008427 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.044366 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.052408 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.157243 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjmm5\" (UniqueName: \"kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5\") pod \"ad2223ee-aa7f-434c-9f0d-8da7c8071f48\" (UID: \"ad2223ee-aa7f-434c-9f0d-8da7c8071f48\") " Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.162069 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5" (OuterVolumeSpecName: "kube-api-access-tjmm5") pod "ad2223ee-aa7f-434c-9f0d-8da7c8071f48" (UID: "ad2223ee-aa7f-434c-9f0d-8da7c8071f48"). InnerVolumeSpecName "kube-api-access-tjmm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.259680 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjmm5\" (UniqueName: \"kubernetes.io/projected/ad2223ee-aa7f-434c-9f0d-8da7c8071f48-kube-api-access-tjmm5\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.493330 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 03 14:54:24 crc kubenswrapper[4959]: E1003 14:54:24.494041 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2223ee-aa7f-434c-9f0d-8da7c8071f48" containerName="mariadb-client-1" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.494060 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2223ee-aa7f-434c-9f0d-8da7c8071f48" containerName="mariadb-client-1" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.494428 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2223ee-aa7f-434c-9f0d-8da7c8071f48" containerName="mariadb-client-1" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.496024 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.509115 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.563271 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgqhh\" (UniqueName: \"kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh\") pod \"mariadb-client-4-default\" (UID: \"e6e5eef2-8768-4a65-a440-3b54f16bbc3b\") " pod="openstack/mariadb-client-4-default" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.569635 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe0f6aad187520c6018b0dd4798e7988bc3128e01e5e2b2e86ded2b8b32e2a02" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.569694 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.666610 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgqhh\" (UniqueName: \"kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh\") pod \"mariadb-client-4-default\" (UID: \"e6e5eef2-8768-4a65-a440-3b54f16bbc3b\") " pod="openstack/mariadb-client-4-default" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.685225 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgqhh\" (UniqueName: \"kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh\") pod \"mariadb-client-4-default\" (UID: \"e6e5eef2-8768-4a65-a440-3b54f16bbc3b\") " pod="openstack/mariadb-client-4-default" Oct 03 14:54:24 crc kubenswrapper[4959]: I1003 14:54:24.819337 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 03 14:54:25 crc kubenswrapper[4959]: I1003 14:54:25.475974 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 03 14:54:25 crc kubenswrapper[4959]: W1003 14:54:25.477766 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6e5eef2_8768_4a65_a440_3b54f16bbc3b.slice/crio-0b346b0df918488d3f3faa2807d74b226959911967325343d3bf436a97f136b0 WatchSource:0}: Error finding container 0b346b0df918488d3f3faa2807d74b226959911967325343d3bf436a97f136b0: Status 404 returned error can't find the container with id 0b346b0df918488d3f3faa2807d74b226959911967325343d3bf436a97f136b0 Oct 03 14:54:25 crc kubenswrapper[4959]: I1003 14:54:25.578302 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"e6e5eef2-8768-4a65-a440-3b54f16bbc3b","Type":"ContainerStarted","Data":"0b346b0df918488d3f3faa2807d74b226959911967325343d3bf436a97f136b0"} Oct 03 14:54:25 crc kubenswrapper[4959]: I1003 14:54:25.697362 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad2223ee-aa7f-434c-9f0d-8da7c8071f48" path="/var/lib/kubelet/pods/ad2223ee-aa7f-434c-9f0d-8da7c8071f48/volumes" Oct 03 14:54:26 crc kubenswrapper[4959]: I1003 14:54:26.590522 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6e5eef2-8768-4a65-a440-3b54f16bbc3b" containerID="c27b4b059704e1d5eafa57548fb8ce9134fb5032cd5ddc79d8a73b05c87c88ee" exitCode=0 Oct 03 14:54:26 crc kubenswrapper[4959]: I1003 14:54:26.590651 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"e6e5eef2-8768-4a65-a440-3b54f16bbc3b","Type":"ContainerDied","Data":"c27b4b059704e1d5eafa57548fb8ce9134fb5032cd5ddc79d8a73b05c87c88ee"} Oct 03 14:54:27 crc kubenswrapper[4959]: I1003 14:54:27.990811 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.012425 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_e6e5eef2-8768-4a65-a440-3b54f16bbc3b/mariadb-client-4-default/0.log" Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.030820 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgqhh\" (UniqueName: \"kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh\") pod \"e6e5eef2-8768-4a65-a440-3b54f16bbc3b\" (UID: \"e6e5eef2-8768-4a65-a440-3b54f16bbc3b\") " Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.038846 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.040421 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh" (OuterVolumeSpecName: "kube-api-access-cgqhh") pod "e6e5eef2-8768-4a65-a440-3b54f16bbc3b" (UID: "e6e5eef2-8768-4a65-a440-3b54f16bbc3b"). InnerVolumeSpecName "kube-api-access-cgqhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.045548 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.132649 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgqhh\" (UniqueName: \"kubernetes.io/projected/e6e5eef2-8768-4a65-a440-3b54f16bbc3b-kube-api-access-cgqhh\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.610972 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b346b0df918488d3f3faa2807d74b226959911967325343d3bf436a97f136b0" Oct 03 14:54:28 crc kubenswrapper[4959]: I1003 14:54:28.611037 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 03 14:54:29 crc kubenswrapper[4959]: I1003 14:54:29.699541 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e5eef2-8768-4a65-a440-3b54f16bbc3b" path="/var/lib/kubelet/pods/e6e5eef2-8768-4a65-a440-3b54f16bbc3b/volumes" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.319774 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 03 14:54:32 crc kubenswrapper[4959]: E1003 14:54:32.322049 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e5eef2-8768-4a65-a440-3b54f16bbc3b" containerName="mariadb-client-4-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.322089 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e5eef2-8768-4a65-a440-3b54f16bbc3b" containerName="mariadb-client-4-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.322335 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e5eef2-8768-4a65-a440-3b54f16bbc3b" containerName="mariadb-client-4-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.322929 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.328330 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vg28g" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.335543 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.507316 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzh9\" (UniqueName: \"kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9\") pod \"mariadb-client-5-default\" (UID: \"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b\") " pod="openstack/mariadb-client-5-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.609135 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzh9\" (UniqueName: \"kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9\") pod \"mariadb-client-5-default\" (UID: \"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b\") " pod="openstack/mariadb-client-5-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.643279 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzh9\" (UniqueName: \"kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9\") pod \"mariadb-client-5-default\" (UID: \"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b\") " pod="openstack/mariadb-client-5-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.651178 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 03 14:54:32 crc kubenswrapper[4959]: I1003 14:54:32.686033 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:54:32 crc kubenswrapper[4959]: E1003 14:54:32.686398 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:54:33 crc kubenswrapper[4959]: I1003 14:54:33.180070 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 03 14:54:33 crc kubenswrapper[4959]: I1003 14:54:33.657207 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b","Type":"ContainerStarted","Data":"c43502030b2aae4cd3d9ef4a7497888cfc891f84e054e895b80c344f783dda10"} Oct 03 14:54:34 crc kubenswrapper[4959]: I1003 14:54:34.671819 4959 generic.go:334] "Generic (PLEG): container finished" podID="86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" containerID="95aeaa7374f1b9d3ac416e7141ffce96300d1bcdac5bed3265593e5e06f694a3" exitCode=0 Oct 03 14:54:34 crc kubenswrapper[4959]: I1003 14:54:34.672051 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b","Type":"ContainerDied","Data":"95aeaa7374f1b9d3ac416e7141ffce96300d1bcdac5bed3265593e5e06f694a3"} Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.107794 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.127874 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_86b9a7ff-0cab-4ae4-9057-e8c7a646d69b/mariadb-client-5-default/0.log" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.159785 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.174011 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.282370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glzh9\" (UniqueName: \"kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9\") pod \"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b\" (UID: \"86b9a7ff-0cab-4ae4-9057-e8c7a646d69b\") " Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.291111 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9" (OuterVolumeSpecName: "kube-api-access-glzh9") pod "86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" (UID: "86b9a7ff-0cab-4ae4-9057-e8c7a646d69b"). InnerVolumeSpecName "kube-api-access-glzh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.332285 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 03 14:54:36 crc kubenswrapper[4959]: E1003 14:54:36.332885 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" containerName="mariadb-client-5-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.332917 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" containerName="mariadb-client-5-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.333176 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" containerName="mariadb-client-5-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.334016 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.358754 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.384579 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glzh9\" (UniqueName: \"kubernetes.io/projected/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b-kube-api-access-glzh9\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.486825 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hljlq\" (UniqueName: \"kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq\") pod \"mariadb-client-6-default\" (UID: \"4fd57da1-bb97-43f6-a1bd-588033ba1fb5\") " pod="openstack/mariadb-client-6-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.588836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hljlq\" (UniqueName: \"kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq\") pod \"mariadb-client-6-default\" (UID: \"4fd57da1-bb97-43f6-a1bd-588033ba1fb5\") " pod="openstack/mariadb-client-6-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.697116 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c43502030b2aae4cd3d9ef4a7497888cfc891f84e054e895b80c344f783dda10" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.697245 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 03 14:54:36 crc kubenswrapper[4959]: I1003 14:54:36.752842 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hljlq\" (UniqueName: \"kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq\") pod \"mariadb-client-6-default\" (UID: \"4fd57da1-bb97-43f6-a1bd-588033ba1fb5\") " pod="openstack/mariadb-client-6-default" Oct 03 14:54:37 crc kubenswrapper[4959]: I1003 14:54:37.029501 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 03 14:54:37 crc kubenswrapper[4959]: I1003 14:54:37.594845 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 03 14:54:37 crc kubenswrapper[4959]: I1003 14:54:37.695898 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b9a7ff-0cab-4ae4-9057-e8c7a646d69b" path="/var/lib/kubelet/pods/86b9a7ff-0cab-4ae4-9057-e8c7a646d69b/volumes" Oct 03 14:54:37 crc kubenswrapper[4959]: I1003 14:54:37.707134 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"4fd57da1-bb97-43f6-a1bd-588033ba1fb5","Type":"ContainerStarted","Data":"197c8cec989868afca7365034cea1cb1631ef26fe80dcb19b35443fa6fafeb56"} Oct 03 14:54:38 crc kubenswrapper[4959]: I1003 14:54:38.723060 4959 generic.go:334] "Generic (PLEG): container finished" podID="4fd57da1-bb97-43f6-a1bd-588033ba1fb5" containerID="37d2db0004bd25d622ca66d7bedcae1b058d8879f4ea407e1188921432b60a3e" exitCode=0 Oct 03 14:54:38 crc kubenswrapper[4959]: I1003 14:54:38.723248 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"4fd57da1-bb97-43f6-a1bd-588033ba1fb5","Type":"ContainerDied","Data":"37d2db0004bd25d622ca66d7bedcae1b058d8879f4ea407e1188921432b60a3e"} Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.138212 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.219435 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_4fd57da1-bb97-43f6-a1bd-588033ba1fb5/mariadb-client-6-default/0.log" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.254943 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hljlq\" (UniqueName: \"kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq\") pod \"4fd57da1-bb97-43f6-a1bd-588033ba1fb5\" (UID: \"4fd57da1-bb97-43f6-a1bd-588033ba1fb5\") " Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.259976 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.260593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq" (OuterVolumeSpecName: "kube-api-access-hljlq") pod "4fd57da1-bb97-43f6-a1bd-588033ba1fb5" (UID: "4fd57da1-bb97-43f6-a1bd-588033ba1fb5"). InnerVolumeSpecName "kube-api-access-hljlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.265521 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.357113 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hljlq\" (UniqueName: \"kubernetes.io/projected/4fd57da1-bb97-43f6-a1bd-588033ba1fb5-kube-api-access-hljlq\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.465505 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 03 14:54:40 crc kubenswrapper[4959]: E1003 14:54:40.466047 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fd57da1-bb97-43f6-a1bd-588033ba1fb5" containerName="mariadb-client-6-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.466085 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fd57da1-bb97-43f6-a1bd-588033ba1fb5" containerName="mariadb-client-6-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.466395 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fd57da1-bb97-43f6-a1bd-588033ba1fb5" containerName="mariadb-client-6-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.467051 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.474827 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.662871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gz6r\" (UniqueName: \"kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r\") pod \"mariadb-client-7-default\" (UID: \"40b7dddf-9a91-4164-a633-ef731cbffef1\") " pod="openstack/mariadb-client-7-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.741310 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="197c8cec989868afca7365034cea1cb1631ef26fe80dcb19b35443fa6fafeb56" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.741362 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.765134 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gz6r\" (UniqueName: \"kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r\") pod \"mariadb-client-7-default\" (UID: \"40b7dddf-9a91-4164-a633-ef731cbffef1\") " pod="openstack/mariadb-client-7-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.794303 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gz6r\" (UniqueName: \"kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r\") pod \"mariadb-client-7-default\" (UID: \"40b7dddf-9a91-4164-a633-ef731cbffef1\") " pod="openstack/mariadb-client-7-default" Oct 03 14:54:40 crc kubenswrapper[4959]: I1003 14:54:40.794870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 03 14:54:41 crc kubenswrapper[4959]: I1003 14:54:41.180038 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 03 14:54:41 crc kubenswrapper[4959]: I1003 14:54:41.705377 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fd57da1-bb97-43f6-a1bd-588033ba1fb5" path="/var/lib/kubelet/pods/4fd57da1-bb97-43f6-a1bd-588033ba1fb5/volumes" Oct 03 14:54:41 crc kubenswrapper[4959]: I1003 14:54:41.753579 4959 generic.go:334] "Generic (PLEG): container finished" podID="40b7dddf-9a91-4164-a633-ef731cbffef1" containerID="d45abb6a8cc72c7299bd7d562274cbb9ffd11a8079995111da94a9a7a217048b" exitCode=0 Oct 03 14:54:41 crc kubenswrapper[4959]: I1003 14:54:41.753649 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"40b7dddf-9a91-4164-a633-ef731cbffef1","Type":"ContainerDied","Data":"d45abb6a8cc72c7299bd7d562274cbb9ffd11a8079995111da94a9a7a217048b"} Oct 03 14:54:41 crc kubenswrapper[4959]: I1003 14:54:41.753688 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"40b7dddf-9a91-4164-a633-ef731cbffef1","Type":"ContainerStarted","Data":"7dad3096a2ce9e297bc932ae3a066ed380f3fbb1283461f891b4f39bebab4826"} Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.196310 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.213252 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_40b7dddf-9a91-4164-a633-ef731cbffef1/mariadb-client-7-default/0.log" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.244761 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.255390 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.312346 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gz6r\" (UniqueName: \"kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r\") pod \"40b7dddf-9a91-4164-a633-ef731cbffef1\" (UID: \"40b7dddf-9a91-4164-a633-ef731cbffef1\") " Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.317679 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r" (OuterVolumeSpecName: "kube-api-access-6gz6r") pod "40b7dddf-9a91-4164-a633-ef731cbffef1" (UID: "40b7dddf-9a91-4164-a633-ef731cbffef1"). InnerVolumeSpecName "kube-api-access-6gz6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.385323 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 03 14:54:43 crc kubenswrapper[4959]: E1003 14:54:43.385671 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b7dddf-9a91-4164-a633-ef731cbffef1" containerName="mariadb-client-7-default" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.385686 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b7dddf-9a91-4164-a633-ef731cbffef1" containerName="mariadb-client-7-default" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.385838 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b7dddf-9a91-4164-a633-ef731cbffef1" containerName="mariadb-client-7-default" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.386487 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.402544 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.415249 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgxjp\" (UniqueName: \"kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp\") pod \"mariadb-client-2\" (UID: \"818e443b-8175-4028-b1b1-a1e34e4e03bb\") " pod="openstack/mariadb-client-2" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.415321 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gz6r\" (UniqueName: \"kubernetes.io/projected/40b7dddf-9a91-4164-a633-ef731cbffef1-kube-api-access-6gz6r\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.516050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgxjp\" (UniqueName: \"kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp\") pod \"mariadb-client-2\" (UID: \"818e443b-8175-4028-b1b1-a1e34e4e03bb\") " pod="openstack/mariadb-client-2" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.536069 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgxjp\" (UniqueName: \"kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp\") pod \"mariadb-client-2\" (UID: \"818e443b-8175-4028-b1b1-a1e34e4e03bb\") " pod="openstack/mariadb-client-2" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.706148 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.708724 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b7dddf-9a91-4164-a633-ef731cbffef1" path="/var/lib/kubelet/pods/40b7dddf-9a91-4164-a633-ef731cbffef1/volumes" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.777766 4959 scope.go:117] "RemoveContainer" containerID="d45abb6a8cc72c7299bd7d562274cbb9ffd11a8079995111da94a9a7a217048b" Oct 03 14:54:43 crc kubenswrapper[4959]: I1003 14:54:43.777912 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 03 14:54:44 crc kubenswrapper[4959]: I1003 14:54:44.234694 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 03 14:54:44 crc kubenswrapper[4959]: I1003 14:54:44.790501 4959 generic.go:334] "Generic (PLEG): container finished" podID="818e443b-8175-4028-b1b1-a1e34e4e03bb" containerID="591346f15def5fc51a579ec324f7321064125f246861471bb162fdffe85b79b2" exitCode=0 Oct 03 14:54:44 crc kubenswrapper[4959]: I1003 14:54:44.790612 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"818e443b-8175-4028-b1b1-a1e34e4e03bb","Type":"ContainerDied","Data":"591346f15def5fc51a579ec324f7321064125f246861471bb162fdffe85b79b2"} Oct 03 14:54:44 crc kubenswrapper[4959]: I1003 14:54:44.790704 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"818e443b-8175-4028-b1b1-a1e34e4e03bb","Type":"ContainerStarted","Data":"aa791b61a082475fc9985c7338cce1c71221ac6c8f813e566251368d9fb9b278"} Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.193370 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.215637 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_818e443b-8175-4028-b1b1-a1e34e4e03bb/mariadb-client-2/0.log" Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.241529 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.247910 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.262461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgxjp\" (UniqueName: \"kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp\") pod \"818e443b-8175-4028-b1b1-a1e34e4e03bb\" (UID: \"818e443b-8175-4028-b1b1-a1e34e4e03bb\") " Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.270571 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp" (OuterVolumeSpecName: "kube-api-access-hgxjp") pod "818e443b-8175-4028-b1b1-a1e34e4e03bb" (UID: "818e443b-8175-4028-b1b1-a1e34e4e03bb"). InnerVolumeSpecName "kube-api-access-hgxjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.364113 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgxjp\" (UniqueName: \"kubernetes.io/projected/818e443b-8175-4028-b1b1-a1e34e4e03bb-kube-api-access-hgxjp\") on node \"crc\" DevicePath \"\"" Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.813132 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa791b61a082475fc9985c7338cce1c71221ac6c8f813e566251368d9fb9b278" Oct 03 14:54:46 crc kubenswrapper[4959]: I1003 14:54:46.813295 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 03 14:54:47 crc kubenswrapper[4959]: I1003 14:54:47.688653 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:54:47 crc kubenswrapper[4959]: E1003 14:54:47.689440 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:54:47 crc kubenswrapper[4959]: I1003 14:54:47.704781 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="818e443b-8175-4028-b1b1-a1e34e4e03bb" path="/var/lib/kubelet/pods/818e443b-8175-4028-b1b1-a1e34e4e03bb/volumes" Oct 03 14:54:59 crc kubenswrapper[4959]: I1003 14:54:59.685827 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:54:59 crc kubenswrapper[4959]: E1003 14:54:59.686653 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:55:05 crc kubenswrapper[4959]: I1003 14:55:05.041030 4959 scope.go:117] "RemoveContainer" containerID="260d3d2b16636dd6f0ba6d7d7879202aebd1b1bcf5b558083c9496350c2d54c9" Oct 03 14:55:11 crc kubenswrapper[4959]: I1003 14:55:11.693152 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:55:11 crc kubenswrapper[4959]: E1003 14:55:11.694393 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:55:23 crc kubenswrapper[4959]: I1003 14:55:23.686447 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:55:23 crc kubenswrapper[4959]: E1003 14:55:23.687508 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:55:35 crc kubenswrapper[4959]: I1003 14:55:35.685347 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:55:35 crc kubenswrapper[4959]: E1003 14:55:35.686121 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:55:50 crc kubenswrapper[4959]: I1003 14:55:50.685379 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:55:50 crc kubenswrapper[4959]: E1003 14:55:50.686284 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:56:04 crc kubenswrapper[4959]: I1003 14:56:04.686480 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:56:04 crc kubenswrapper[4959]: E1003 14:56:04.687315 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:56:16 crc kubenswrapper[4959]: I1003 14:56:16.685153 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:56:16 crc kubenswrapper[4959]: E1003 14:56:16.685956 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:56:27 crc kubenswrapper[4959]: I1003 14:56:27.685442 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:56:27 crc kubenswrapper[4959]: E1003 14:56:27.686019 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:56:39 crc kubenswrapper[4959]: I1003 14:56:39.685299 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:56:39 crc kubenswrapper[4959]: E1003 14:56:39.686304 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:56:51 crc kubenswrapper[4959]: I1003 14:56:51.694710 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:56:51 crc kubenswrapper[4959]: E1003 14:56:51.695697 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:57:04 crc kubenswrapper[4959]: I1003 14:57:04.686776 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:57:04 crc kubenswrapper[4959]: E1003 14:57:04.687586 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:57:18 crc kubenswrapper[4959]: I1003 14:57:18.685908 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:57:18 crc kubenswrapper[4959]: E1003 14:57:18.687965 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:57:29 crc kubenswrapper[4959]: I1003 14:57:29.686359 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:57:29 crc kubenswrapper[4959]: E1003 14:57:29.687409 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:57:40 crc kubenswrapper[4959]: I1003 14:57:40.686057 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:57:40 crc kubenswrapper[4959]: E1003 14:57:40.686899 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:57:54 crc kubenswrapper[4959]: I1003 14:57:54.686156 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:57:54 crc kubenswrapper[4959]: E1003 14:57:54.687443 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:58:06 crc kubenswrapper[4959]: I1003 14:58:06.686563 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:58:06 crc kubenswrapper[4959]: E1003 14:58:06.689488 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:58:18 crc kubenswrapper[4959]: I1003 14:58:18.686602 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:58:18 crc kubenswrapper[4959]: E1003 14:58:18.687665 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:58:29 crc kubenswrapper[4959]: I1003 14:58:29.685919 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:58:29 crc kubenswrapper[4959]: E1003 14:58:29.686975 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 14:58:40 crc kubenswrapper[4959]: I1003 14:58:40.685584 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 14:58:40 crc kubenswrapper[4959]: I1003 14:58:40.964034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e"} Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.141633 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 03 14:58:58 crc kubenswrapper[4959]: E1003 14:58:58.142311 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="818e443b-8175-4028-b1b1-a1e34e4e03bb" containerName="mariadb-client-2" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.142323 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="818e443b-8175-4028-b1b1-a1e34e4e03bb" containerName="mariadb-client-2" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.142463 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="818e443b-8175-4028-b1b1-a1e34e4e03bb" containerName="mariadb-client-2" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.144313 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.146893 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vg28g" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.160565 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.333030 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8c890935-40c0-4a61-90b7-785d3225978b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c890935-40c0-4a61-90b7-785d3225978b\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.333282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgbr2\" (UniqueName: \"kubernetes.io/projected/19472efc-08e5-4271-ada9-a7664d2dcef6-kube-api-access-bgbr2\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.435552 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgbr2\" (UniqueName: \"kubernetes.io/projected/19472efc-08e5-4271-ada9-a7664d2dcef6-kube-api-access-bgbr2\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.435701 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8c890935-40c0-4a61-90b7-785d3225978b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c890935-40c0-4a61-90b7-785d3225978b\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.440269 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.440334 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8c890935-40c0-4a61-90b7-785d3225978b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c890935-40c0-4a61-90b7-785d3225978b\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4e8bf5b4f5a6a811e6d86dd006dcc9936eb4141b3008ce4a68450bb0da70c9fa/globalmount\"" pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.478151 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgbr2\" (UniqueName: \"kubernetes.io/projected/19472efc-08e5-4271-ada9-a7664d2dcef6-kube-api-access-bgbr2\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.487026 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8c890935-40c0-4a61-90b7-785d3225978b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c890935-40c0-4a61-90b7-785d3225978b\") pod \"mariadb-copy-data\" (UID: \"19472efc-08e5-4271-ada9-a7664d2dcef6\") " pod="openstack/mariadb-copy-data" Oct 03 14:58:58 crc kubenswrapper[4959]: I1003 14:58:58.771320 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 03 14:58:59 crc kubenswrapper[4959]: I1003 14:58:59.350142 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 03 14:58:59 crc kubenswrapper[4959]: W1003 14:58:59.357271 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19472efc_08e5_4271_ada9_a7664d2dcef6.slice/crio-1809cea1738e74ae64333cd8eb4121dc1d1bbaca64f7d99a2a910e358aa5eee5 WatchSource:0}: Error finding container 1809cea1738e74ae64333cd8eb4121dc1d1bbaca64f7d99a2a910e358aa5eee5: Status 404 returned error can't find the container with id 1809cea1738e74ae64333cd8eb4121dc1d1bbaca64f7d99a2a910e358aa5eee5 Oct 03 14:59:00 crc kubenswrapper[4959]: I1003 14:59:00.111229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"19472efc-08e5-4271-ada9-a7664d2dcef6","Type":"ContainerStarted","Data":"d7bdeee309fae332e2cc467382c21e550ac718a78ac0b99a4851a8d527be647e"} Oct 03 14:59:00 crc kubenswrapper[4959]: I1003 14:59:00.111273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"19472efc-08e5-4271-ada9-a7664d2dcef6","Type":"ContainerStarted","Data":"1809cea1738e74ae64333cd8eb4121dc1d1bbaca64f7d99a2a910e358aa5eee5"} Oct 03 14:59:00 crc kubenswrapper[4959]: I1003 14:59:00.133850 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.1338293999999998 podStartE2EDuration="3.1338294s" podCreationTimestamp="2025-10-03 14:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:00.127608717 +0000 UTC m=+5309.330952144" watchObservedRunningTime="2025-10-03 14:59:00.1338294 +0000 UTC m=+5309.337172817" Oct 03 14:59:01 crc kubenswrapper[4959]: I1003 14:59:01.839390 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:01 crc kubenswrapper[4959]: I1003 14:59:01.840718 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:01 crc kubenswrapper[4959]: I1003 14:59:01.851039 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:01 crc kubenswrapper[4959]: I1003 14:59:01.990412 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g49zt\" (UniqueName: \"kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt\") pod \"mariadb-client\" (UID: \"a389f4e5-a6e1-4a43-bb3a-a346b60ce269\") " pod="openstack/mariadb-client" Oct 03 14:59:02 crc kubenswrapper[4959]: I1003 14:59:02.091808 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g49zt\" (UniqueName: \"kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt\") pod \"mariadb-client\" (UID: \"a389f4e5-a6e1-4a43-bb3a-a346b60ce269\") " pod="openstack/mariadb-client" Oct 03 14:59:02 crc kubenswrapper[4959]: I1003 14:59:02.121282 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g49zt\" (UniqueName: \"kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt\") pod \"mariadb-client\" (UID: \"a389f4e5-a6e1-4a43-bb3a-a346b60ce269\") " pod="openstack/mariadb-client" Oct 03 14:59:02 crc kubenswrapper[4959]: I1003 14:59:02.176564 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:02 crc kubenswrapper[4959]: I1003 14:59:02.608499 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:03 crc kubenswrapper[4959]: I1003 14:59:03.145787 4959 generic.go:334] "Generic (PLEG): container finished" podID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" containerID="2f8d800d6450c81e3af1f9ec6e909425c134f05dc3094f6f0ef288ed781edcf3" exitCode=0 Oct 03 14:59:03 crc kubenswrapper[4959]: I1003 14:59:03.147673 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a389f4e5-a6e1-4a43-bb3a-a346b60ce269","Type":"ContainerDied","Data":"2f8d800d6450c81e3af1f9ec6e909425c134f05dc3094f6f0ef288ed781edcf3"} Oct 03 14:59:03 crc kubenswrapper[4959]: I1003 14:59:03.147856 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a389f4e5-a6e1-4a43-bb3a-a346b60ce269","Type":"ContainerStarted","Data":"5cf3ded878819cd64b66c3d4b3da2d14a6e4525cc8201ef3e171ce6793f543c2"} Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.442611 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.462817 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_a389f4e5-a6e1-4a43-bb3a-a346b60ce269/mariadb-client/0.log" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.493583 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.498268 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.530730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g49zt\" (UniqueName: \"kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt\") pod \"a389f4e5-a6e1-4a43-bb3a-a346b60ce269\" (UID: \"a389f4e5-a6e1-4a43-bb3a-a346b60ce269\") " Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.539572 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt" (OuterVolumeSpecName: "kube-api-access-g49zt") pod "a389f4e5-a6e1-4a43-bb3a-a346b60ce269" (UID: "a389f4e5-a6e1-4a43-bb3a-a346b60ce269"). InnerVolumeSpecName "kube-api-access-g49zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.633393 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g49zt\" (UniqueName: \"kubernetes.io/projected/a389f4e5-a6e1-4a43-bb3a-a346b60ce269-kube-api-access-g49zt\") on node \"crc\" DevicePath \"\"" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.696807 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:04 crc kubenswrapper[4959]: E1003 14:59:04.697401 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" containerName="mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.697471 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" containerName="mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.697688 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" containerName="mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.698284 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.708864 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.836686 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt5nb\" (UniqueName: \"kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb\") pod \"mariadb-client\" (UID: \"bab6c4ed-2063-4e9e-9704-9d5755d32a64\") " pod="openstack/mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.939927 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt5nb\" (UniqueName: \"kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb\") pod \"mariadb-client\" (UID: \"bab6c4ed-2063-4e9e-9704-9d5755d32a64\") " pod="openstack/mariadb-client" Oct 03 14:59:04 crc kubenswrapper[4959]: I1003 14:59:04.961737 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt5nb\" (UniqueName: \"kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb\") pod \"mariadb-client\" (UID: \"bab6c4ed-2063-4e9e-9704-9d5755d32a64\") " pod="openstack/mariadb-client" Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.023613 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.161987 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cf3ded878819cd64b66c3d4b3da2d14a6e4525cc8201ef3e171ce6793f543c2" Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.162054 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.184144 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" podUID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.460084 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:05 crc kubenswrapper[4959]: W1003 14:59:05.464296 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbab6c4ed_2063_4e9e_9704_9d5755d32a64.slice/crio-ab0980b4dcc1cfa4f1a1ac52f930ad92420727ed3ea50b34a5cc3fdd96cefecb WatchSource:0}: Error finding container ab0980b4dcc1cfa4f1a1ac52f930ad92420727ed3ea50b34a5cc3fdd96cefecb: Status 404 returned error can't find the container with id ab0980b4dcc1cfa4f1a1ac52f930ad92420727ed3ea50b34a5cc3fdd96cefecb Oct 03 14:59:05 crc kubenswrapper[4959]: I1003 14:59:05.703315 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a389f4e5-a6e1-4a43-bb3a-a346b60ce269" path="/var/lib/kubelet/pods/a389f4e5-a6e1-4a43-bb3a-a346b60ce269/volumes" Oct 03 14:59:06 crc kubenswrapper[4959]: I1003 14:59:06.170830 4959 generic.go:334] "Generic (PLEG): container finished" podID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" containerID="921a44df34fd2b56735ad3922bfec8a49ce4d559c1d9920f476f096c6b14a82b" exitCode=0 Oct 03 14:59:06 crc kubenswrapper[4959]: I1003 14:59:06.170879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bab6c4ed-2063-4e9e-9704-9d5755d32a64","Type":"ContainerDied","Data":"921a44df34fd2b56735ad3922bfec8a49ce4d559c1d9920f476f096c6b14a82b"} Oct 03 14:59:06 crc kubenswrapper[4959]: I1003 14:59:06.170905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"bab6c4ed-2063-4e9e-9704-9d5755d32a64","Type":"ContainerStarted","Data":"ab0980b4dcc1cfa4f1a1ac52f930ad92420727ed3ea50b34a5cc3fdd96cefecb"} Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.589017 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.604778 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_bab6c4ed-2063-4e9e-9704-9d5755d32a64/mariadb-client/0.log" Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.629405 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.634250 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.686124 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt5nb\" (UniqueName: \"kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb\") pod \"bab6c4ed-2063-4e9e-9704-9d5755d32a64\" (UID: \"bab6c4ed-2063-4e9e-9704-9d5755d32a64\") " Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.695807 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb" (OuterVolumeSpecName: "kube-api-access-dt5nb") pod "bab6c4ed-2063-4e9e-9704-9d5755d32a64" (UID: "bab6c4ed-2063-4e9e-9704-9d5755d32a64"). InnerVolumeSpecName "kube-api-access-dt5nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:59:07 crc kubenswrapper[4959]: I1003 14:59:07.787979 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt5nb\" (UniqueName: \"kubernetes.io/projected/bab6c4ed-2063-4e9e-9704-9d5755d32a64-kube-api-access-dt5nb\") on node \"crc\" DevicePath \"\"" Oct 03 14:59:08 crc kubenswrapper[4959]: I1003 14:59:08.190213 4959 scope.go:117] "RemoveContainer" containerID="921a44df34fd2b56735ad3922bfec8a49ce4d559c1d9920f476f096c6b14a82b" Oct 03 14:59:08 crc kubenswrapper[4959]: I1003 14:59:08.190279 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 03 14:59:09 crc kubenswrapper[4959]: I1003 14:59:09.696466 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" path="/var/lib/kubelet/pods/bab6c4ed-2063-4e9e-9704-9d5755d32a64/volumes" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.064772 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 14:59:52 crc kubenswrapper[4959]: E1003 14:59:52.066724 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" containerName="mariadb-client" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.066807 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" containerName="mariadb-client" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.067047 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab6c4ed-2063-4e9e-9704-9d5755d32a64" containerName="mariadb-client" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.068382 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.070634 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.071359 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.071649 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-x6psd" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.075095 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.120429 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.122324 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.129819 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.133068 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.144347 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.144622 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.144658 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.144912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.144972 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.145174 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mjsv\" (UniqueName: \"kubernetes.io/projected/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-kube-api-access-6mjsv\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.156883 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.170678 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247221 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247299 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dc2579-070a-44a4-8063-89979bc14f52-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247326 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl65f\" (UniqueName: \"kubernetes.io/projected/1ad50a25-04e2-433b-af7d-1d73ddf151a9-kube-api-access-fl65f\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247348 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad50a25-04e2-433b-af7d-1d73ddf151a9-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247409 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247437 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247478 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.247498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ad50a25-04e2-433b-af7d-1d73ddf151a9-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248392 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248435 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248504 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mjsv\" (UniqueName: \"kubernetes.io/projected/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-kube-api-access-6mjsv\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248536 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-config\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248624 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248752 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx4qt\" (UniqueName: \"kubernetes.io/projected/76dc2579-070a-44a4-8063-89979bc14f52-kube-api-access-bx4qt\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248782 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248851 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248875 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-config\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.248895 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76dc2579-070a-44a4-8063-89979bc14f52-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.250122 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.251796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.253739 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.253768 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/df67df603a873eb10abc0f3f9b64bb14788caf8872359bea560a4a9eddb2325f/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.268037 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.268059 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mjsv\" (UniqueName: \"kubernetes.io/projected/b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5-kube-api-access-6mjsv\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.271718 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.275465 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.278078 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-5ltjk" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.279694 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.279931 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.279984 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.297532 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.305045 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.308488 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.309801 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.320358 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.336898 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.341537 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf06ea3b-b6ca-41e6-b5fc-3f2ced34152b\") pod \"ovsdbserver-nb-0\" (UID: \"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5\") " pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.350905 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl65f\" (UniqueName: \"kubernetes.io/projected/1ad50a25-04e2-433b-af7d-1d73ddf151a9-kube-api-access-fl65f\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.351137 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad50a25-04e2-433b-af7d-1d73ddf151a9-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.351250 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/053f4c08-15f4-4adb-8b32-e3e200fd4197-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.351332 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.351703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jmgk\" (UniqueName: \"kubernetes.io/projected/97c621eb-b91e-42fb-91df-25cb1782d264-kube-api-access-8jmgk\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352272 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ad50a25-04e2-433b-af7d-1d73ddf151a9-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352368 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352440 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-config\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352509 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-config\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352582 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgwc5\" (UniqueName: \"kubernetes.io/projected/053f4c08-15f4-4adb-8b32-e3e200fd4197-kube-api-access-cgwc5\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352659 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p88vl\" (UniqueName: \"kubernetes.io/projected/29b1ce1a-4945-46ba-994b-4cf3167e2070-kube-api-access-p88vl\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352743 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.352948 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b1ce1a-4945-46ba-994b-4cf3167e2070-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.353032 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f4c08-15f4-4adb-8b32-e3e200fd4197-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.353129 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx4qt\" (UniqueName: \"kubernetes.io/projected/76dc2579-070a-44a4-8063-89979bc14f52-kube-api-access-bx4qt\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.353245 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.353327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354448 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354688 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-config\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354833 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354923 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-config\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.354969 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76dc2579-070a-44a4-8063-89979bc14f52-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97c621eb-b91e-42fb-91df-25cb1782d264-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355053 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-config\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355128 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/29b1ce1a-4945-46ba-994b-4cf3167e2070-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355154 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c621eb-b91e-42fb-91df-25cb1782d264-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dc2579-070a-44a4-8063-89979bc14f52-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355439 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355610 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355644 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0e90fa8dd364d30110acb3658f75d55a6e44effe5190da05e74a18760e7f20d9/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.355795 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ad50a25-04e2-433b-af7d-1d73ddf151a9-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.356024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/76dc2579-070a-44a4-8063-89979bc14f52-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.356137 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76dc2579-070a-44a4-8063-89979bc14f52-config\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.357020 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-config\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.360115 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dc2579-070a-44a4-8063-89979bc14f52-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.360720 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.361220 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/29e0c266b331d69a3e2f320611511f27ecb5e27e953cc8ee96f4dafa1066f81f/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.360782 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad50a25-04e2-433b-af7d-1d73ddf151a9-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.362107 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad50a25-04e2-433b-af7d-1d73ddf151a9-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.371633 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl65f\" (UniqueName: \"kubernetes.io/projected/1ad50a25-04e2-433b-af7d-1d73ddf151a9-kube-api-access-fl65f\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.373037 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx4qt\" (UniqueName: \"kubernetes.io/projected/76dc2579-070a-44a4-8063-89979bc14f52-kube-api-access-bx4qt\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.391640 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4f6a66c4-6c96-4d48-b6ca-9088be4e87ec\") pod \"ovsdbserver-nb-2\" (UID: \"76dc2579-070a-44a4-8063-89979bc14f52\") " pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.392726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-48945538-9e2d-45ab-9258-7cc34f45d85d\") pod \"ovsdbserver-nb-1\" (UID: \"1ad50a25-04e2-433b-af7d-1d73ddf151a9\") " pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.399590 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.439704 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457592 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97c621eb-b91e-42fb-91df-25cb1782d264-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457651 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457675 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-config\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/29b1ce1a-4945-46ba-994b-4cf3167e2070-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457740 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c621eb-b91e-42fb-91df-25cb1782d264-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457784 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/053f4c08-15f4-4adb-8b32-e3e200fd4197-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457815 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jmgk\" (UniqueName: \"kubernetes.io/projected/97c621eb-b91e-42fb-91df-25cb1782d264-kube-api-access-8jmgk\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457846 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-config\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457874 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgwc5\" (UniqueName: \"kubernetes.io/projected/053f4c08-15f4-4adb-8b32-e3e200fd4197-kube-api-access-cgwc5\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457900 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p88vl\" (UniqueName: \"kubernetes.io/projected/29b1ce1a-4945-46ba-994b-4cf3167e2070-kube-api-access-p88vl\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457934 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b1ce1a-4945-46ba-994b-4cf3167e2070-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457964 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f4c08-15f4-4adb-8b32-e3e200fd4197-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.457993 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.458017 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.458037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.458057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.458077 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-config\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.458143 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.459031 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/97c621eb-b91e-42fb-91df-25cb1782d264-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.459681 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.460519 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-config\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.460889 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/29b1ce1a-4945-46ba-994b-4cf3167e2070-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.463000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b1ce1a-4945-46ba-994b-4cf3167e2070-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.464682 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.466712 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.466754 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/eccbe46e666b5b10331c8a69cf3e569f57fc57171834e273914b70b8f672b4fb/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.467464 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.467494 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9fd71f25c1fa7bffa44430a5dc90ee7f22bfe940c0baf97a0308d098cbf8f636/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.467887 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.467904 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c9572cd6c6d5c514ad8e09599bb7b3c7d3e0397e7bbf1248467cb0c40fcf52d6/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.468265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29b1ce1a-4945-46ba-994b-4cf3167e2070-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.469422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c621eb-b91e-42fb-91df-25cb1782d264-config\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.470424 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/053f4c08-15f4-4adb-8b32-e3e200fd4197-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.471441 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.473060 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/053f4c08-15f4-4adb-8b32-e3e200fd4197-config\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.477935 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f4c08-15f4-4adb-8b32-e3e200fd4197-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.479465 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97c621eb-b91e-42fb-91df-25cb1782d264-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.482999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p88vl\" (UniqueName: \"kubernetes.io/projected/29b1ce1a-4945-46ba-994b-4cf3167e2070-kube-api-access-p88vl\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.488749 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jmgk\" (UniqueName: \"kubernetes.io/projected/97c621eb-b91e-42fb-91df-25cb1782d264-kube-api-access-8jmgk\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.499035 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgwc5\" (UniqueName: \"kubernetes.io/projected/053f4c08-15f4-4adb-8b32-e3e200fd4197-kube-api-access-cgwc5\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.516840 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-614e17a2-4c71-43c7-a559-70fd6a4ea142\") pod \"ovsdbserver-sb-0\" (UID: \"97c621eb-b91e-42fb-91df-25cb1782d264\") " pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.526380 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d2b6d7ee-c758-42e2-a5de-63dbc14ba268\") pod \"ovsdbserver-sb-2\" (UID: \"29b1ce1a-4945-46ba-994b-4cf3167e2070\") " pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.535999 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a3c18cf9-693c-4944-ac63-d3bfa25d2053\") pod \"ovsdbserver-sb-1\" (UID: \"053f4c08-15f4-4adb-8b32-e3e200fd4197\") " pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.649550 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.769075 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.798959 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:52 crc kubenswrapper[4959]: I1003 14:59:52.977077 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.051266 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 03 14:59:53 crc kubenswrapper[4959]: W1003 14:59:53.156537 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod053f4c08_15f4_4adb_8b32_e3e200fd4197.slice/crio-e89122239031cc0321c598329e9a4ebda856cf4ae3b84510ee2d96ad7237635f WatchSource:0}: Error finding container e89122239031cc0321c598329e9a4ebda856cf4ae3b84510ee2d96ad7237635f: Status 404 returned error can't find the container with id e89122239031cc0321c598329e9a4ebda856cf4ae3b84510ee2d96ad7237635f Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.164250 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.407570 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.603949 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"29b1ce1a-4945-46ba-994b-4cf3167e2070","Type":"ContainerStarted","Data":"9dd99c18937b81b2431fa44208bd7371b060d86ad78fa5e05b17302947fb1a1e"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.604026 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"29b1ce1a-4945-46ba-994b-4cf3167e2070","Type":"ContainerStarted","Data":"b64106e765e3c91a06a87b38b56121ec6b39853966c84aa8c04ebdf85b64e881"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.606184 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"1ad50a25-04e2-433b-af7d-1d73ddf151a9","Type":"ContainerStarted","Data":"2a35135005aeb728f4442e60c7ec5beef00f33e368ecc3ce25727ca647d52627"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.606286 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"1ad50a25-04e2-433b-af7d-1d73ddf151a9","Type":"ContainerStarted","Data":"5da9159b3b99e5244b1a03a52c60cc6f0dc7da7783551727e255b9ead81f1170"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.606303 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"1ad50a25-04e2-433b-af7d-1d73ddf151a9","Type":"ContainerStarted","Data":"40ad5bca4937cb19370e03de0843fa15dd52a3e333438ec53b6fa8cead5ecf34"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.607819 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"053f4c08-15f4-4adb-8b32-e3e200fd4197","Type":"ContainerStarted","Data":"7de346843457f2a20cefdd0b61c80a8b371d1cae5b24bdd19d7bd261914a7243"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.607845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"053f4c08-15f4-4adb-8b32-e3e200fd4197","Type":"ContainerStarted","Data":"02e679a8136a7a200397de36c6df84d758e9f2c35689fa604068a7b4211b278e"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.607867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"053f4c08-15f4-4adb-8b32-e3e200fd4197","Type":"ContainerStarted","Data":"e89122239031cc0321c598329e9a4ebda856cf4ae3b84510ee2d96ad7237635f"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.610710 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5","Type":"ContainerStarted","Data":"8f8b99f664c02e45bbd77d154364170ad47e842c150f46e4ca5cce937aff9b9c"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.610753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5","Type":"ContainerStarted","Data":"7be77b7d2fa6c7f755a16ef3fff82481013973bfcac22dd6c1c56cb43286d609"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.610767 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5","Type":"ContainerStarted","Data":"5589cd5226fc96ff71a9dc422bc6f741ac34b9d9a354d10d956d38e31263430b"} Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.632127 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=2.6320868490000002 podStartE2EDuration="2.632086849s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:53.624827542 +0000 UTC m=+5362.828170979" watchObservedRunningTime="2025-10-03 14:59:53.632086849 +0000 UTC m=+5362.835430276" Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.644703 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.6446845679999997 podStartE2EDuration="2.644684568s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:53.640333431 +0000 UTC m=+5362.843676848" watchObservedRunningTime="2025-10-03 14:59:53.644684568 +0000 UTC m=+5362.848027985" Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.660625 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=2.660605877 podStartE2EDuration="2.660605877s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:53.658581268 +0000 UTC m=+5362.861924685" watchObservedRunningTime="2025-10-03 14:59:53.660605877 +0000 UTC m=+5362.863949294" Oct 03 14:59:53 crc kubenswrapper[4959]: I1003 14:59:53.910706 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 14:59:53 crc kubenswrapper[4959]: W1003 14:59:53.916744 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97c621eb_b91e_42fb_91df_25cb1782d264.slice/crio-882b23aa4e0c8542e47b24c616446e6fa8333c9d6fc6451ca9382d8ce1f28609 WatchSource:0}: Error finding container 882b23aa4e0c8542e47b24c616446e6fa8333c9d6fc6451ca9382d8ce1f28609: Status 404 returned error can't find the container with id 882b23aa4e0c8542e47b24c616446e6fa8333c9d6fc6451ca9382d8ce1f28609 Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.010703 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 03 14:59:54 crc kubenswrapper[4959]: W1003 14:59:54.013629 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76dc2579_070a_44a4_8063_89979bc14f52.slice/crio-98b1f9d308ca636c8db6400a7dbb89183b83c6f72092fb085ef7aabc71b10c8a WatchSource:0}: Error finding container 98b1f9d308ca636c8db6400a7dbb89183b83c6f72092fb085ef7aabc71b10c8a: Status 404 returned error can't find the container with id 98b1f9d308ca636c8db6400a7dbb89183b83c6f72092fb085ef7aabc71b10c8a Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.620658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"29b1ce1a-4945-46ba-994b-4cf3167e2070","Type":"ContainerStarted","Data":"4e55c507a0d36a42e88e979bda90b2d65ccb55e579931b943e2cd506311fc490"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.623342 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"76dc2579-070a-44a4-8063-89979bc14f52","Type":"ContainerStarted","Data":"ac28c3440f0332905b6413a8ce41968e7735233949382bbe1d4a1155fd21ba58"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.623393 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"76dc2579-070a-44a4-8063-89979bc14f52","Type":"ContainerStarted","Data":"512464615a6ce8d22b1c148df2ee3c06026bc3a3b8aaed805403d0c80b019181"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.623410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"76dc2579-070a-44a4-8063-89979bc14f52","Type":"ContainerStarted","Data":"98b1f9d308ca636c8db6400a7dbb89183b83c6f72092fb085ef7aabc71b10c8a"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.626217 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97c621eb-b91e-42fb-91df-25cb1782d264","Type":"ContainerStarted","Data":"040175bad5c0d3afa002d2710d63760000ac666d2db494bc7e43be46d7e0f07a"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.626313 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97c621eb-b91e-42fb-91df-25cb1782d264","Type":"ContainerStarted","Data":"457b8cf912f25414d283eee9a62c48c264ae20164015c94ac73ae88f4f0c4c0e"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.626346 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"97c621eb-b91e-42fb-91df-25cb1782d264","Type":"ContainerStarted","Data":"882b23aa4e0c8542e47b24c616446e6fa8333c9d6fc6451ca9382d8ce1f28609"} Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.663476 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.663446695 podStartE2EDuration="3.663446695s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:54.646746896 +0000 UTC m=+5363.850090343" watchObservedRunningTime="2025-10-03 14:59:54.663446695 +0000 UTC m=+5363.866790132" Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.671572 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.671511513 podStartE2EDuration="3.671511513s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:54.668602962 +0000 UTC m=+5363.871946389" watchObservedRunningTime="2025-10-03 14:59:54.671511513 +0000 UTC m=+5363.874854960" Oct 03 14:59:54 crc kubenswrapper[4959]: I1003 14:59:54.698934 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.698911534 podStartE2EDuration="3.698911534s" podCreationTimestamp="2025-10-03 14:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:59:54.692064696 +0000 UTC m=+5363.895408113" watchObservedRunningTime="2025-10-03 14:59:54.698911534 +0000 UTC m=+5363.902254961" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.399839 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.440903 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.471787 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.650405 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.770686 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:55 crc kubenswrapper[4959]: I1003 14:59:55.799806 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.400346 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.440379 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.472254 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.650478 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.771342 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:57 crc kubenswrapper[4959]: I1003 14:59:57.800229 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.471569 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.518527 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.538102 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.538233 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.584317 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.715788 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.745305 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.746531 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.749515 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.766440 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.809516 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.847616 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.850142 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.893853 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.896876 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.896943 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.897109 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.897169 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc8kv\" (UniqueName: \"kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.998869 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.998932 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.998978 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:58 crc kubenswrapper[4959]: I1003 14:59:58.999004 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc8kv\" (UniqueName: \"kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.000434 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.000837 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.001582 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.015958 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc8kv\" (UniqueName: \"kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv\") pod \"dnsmasq-dns-64848558ff-j2mj5\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.066688 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.159030 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.183992 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.216850 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.217230 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.235785 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.411894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.411957 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpj94\" (UniqueName: \"kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.412042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.412068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.412111 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.514223 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.514312 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.514499 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.514698 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.514746 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpj94\" (UniqueName: \"kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.515173 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.515296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.515889 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.516459 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.535769 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpj94\" (UniqueName: \"kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94\") pod \"dnsmasq-dns-864bc46885-gvsdp\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.552951 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.597510 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 14:59:59 crc kubenswrapper[4959]: I1003 14:59:59.699816 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" event={"ID":"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a","Type":"ContainerStarted","Data":"ad81cae2b4cbc8507db0698c9b50093fc920322db45beac3de27964e42d1b9c0"} Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.040373 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 15:00:00 crc kubenswrapper[4959]: W1003 15:00:00.044034 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda39a3471_a59a_43cf_9e3c_db53b8c0c63e.slice/crio-567b01a34906bfc93d97e03bb9219cbea9aa01e2c2c7318a5e5cccbac5e1807f WatchSource:0}: Error finding container 567b01a34906bfc93d97e03bb9219cbea9aa01e2c2c7318a5e5cccbac5e1807f: Status 404 returned error can't find the container with id 567b01a34906bfc93d97e03bb9219cbea9aa01e2c2c7318a5e5cccbac5e1807f Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.148484 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x"] Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.153854 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.156589 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.156766 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.161628 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x"] Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.328020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.328110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.328170 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2lpt\" (UniqueName: \"kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.430024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.430228 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2lpt\" (UniqueName: \"kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.430462 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.431442 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.433925 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.452624 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2lpt\" (UniqueName: \"kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt\") pod \"collect-profiles-29325060-crq9x\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.530178 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.707553 4959 generic.go:334] "Generic (PLEG): container finished" podID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerID="d3b1f51f3a25001eeb9ef50f61c5abf3df8647ec6e954966c59a0f8cf6ac8af7" exitCode=0 Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.707897 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" event={"ID":"a39a3471-a59a-43cf-9e3c-db53b8c0c63e","Type":"ContainerDied","Data":"d3b1f51f3a25001eeb9ef50f61c5abf3df8647ec6e954966c59a0f8cf6ac8af7"} Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.707946 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" event={"ID":"a39a3471-a59a-43cf-9e3c-db53b8c0c63e","Type":"ContainerStarted","Data":"567b01a34906bfc93d97e03bb9219cbea9aa01e2c2c7318a5e5cccbac5e1807f"} Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.711852 4959 generic.go:334] "Generic (PLEG): container finished" podID="d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" containerID="bee99402c86b02183a2d7065b6f98ebb1f85d454e93e52e5a51ac5260c278719" exitCode=0 Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.711895 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" event={"ID":"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a","Type":"ContainerDied","Data":"bee99402c86b02183a2d7065b6f98ebb1f85d454e93e52e5a51ac5260c278719"} Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.949170 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x"] Oct 03 15:00:00 crc kubenswrapper[4959]: W1003 15:00:00.951397 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f4f4059_389d_4e04_90b3_34ec7a502754.slice/crio-d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c WatchSource:0}: Error finding container d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c: Status 404 returned error can't find the container with id d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c Oct 03 15:00:00 crc kubenswrapper[4959]: I1003 15:00:00.957661 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.141336 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc8kv\" (UniqueName: \"kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv\") pod \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.141694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc\") pod \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.141789 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb\") pod \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.141861 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config\") pod \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\" (UID: \"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a\") " Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.149117 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv" (OuterVolumeSpecName: "kube-api-access-jc8kv") pod "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" (UID: "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a"). InnerVolumeSpecName "kube-api-access-jc8kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.165705 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" (UID: "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.167483 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" (UID: "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.182062 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config" (OuterVolumeSpecName: "config") pod "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" (UID: "d0caa2dc-a0d2-4354-a2d7-05ea981dac5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.243836 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.243874 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.243887 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.243895 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc8kv\" (UniqueName: \"kubernetes.io/projected/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a-kube-api-access-jc8kv\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.723759 4959 generic.go:334] "Generic (PLEG): container finished" podID="6f4f4059-389d-4e04-90b3-34ec7a502754" containerID="ccffb66f52a9bcc89ceb12109241ee53fea8bce5817981398ea259b975a5709e" exitCode=0 Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.723867 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" event={"ID":"6f4f4059-389d-4e04-90b3-34ec7a502754","Type":"ContainerDied","Data":"ccffb66f52a9bcc89ceb12109241ee53fea8bce5817981398ea259b975a5709e"} Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.724111 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" event={"ID":"6f4f4059-389d-4e04-90b3-34ec7a502754","Type":"ContainerStarted","Data":"d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c"} Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.725892 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" event={"ID":"d0caa2dc-a0d2-4354-a2d7-05ea981dac5a","Type":"ContainerDied","Data":"ad81cae2b4cbc8507db0698c9b50093fc920322db45beac3de27964e42d1b9c0"} Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.725933 4959 scope.go:117] "RemoveContainer" containerID="bee99402c86b02183a2d7065b6f98ebb1f85d454e93e52e5a51ac5260c278719" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.726047 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64848558ff-j2mj5" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.730509 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" event={"ID":"a39a3471-a59a-43cf-9e3c-db53b8c0c63e","Type":"ContainerStarted","Data":"8625b7a263056a2ed3fbbe2864e024162970a01f5177078d576ee2ac2f28b288"} Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.730969 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.824300 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.831615 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64848558ff-j2mj5"] Oct 03 15:00:01 crc kubenswrapper[4959]: I1003 15:00:01.836270 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" podStartSLOduration=2.836247557 podStartE2EDuration="2.836247557s" podCreationTimestamp="2025-10-03 14:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:01.810117808 +0000 UTC m=+5371.013461235" watchObservedRunningTime="2025-10-03 15:00:01.836247557 +0000 UTC m=+5371.039590974" Oct 03 15:00:01 crc kubenswrapper[4959]: E1003 15:00:01.844492 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0caa2dc_a0d2_4354_a2d7_05ea981dac5a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0caa2dc_a0d2_4354_a2d7_05ea981dac5a.slice/crio-ad81cae2b4cbc8507db0698c9b50093fc920322db45beac3de27964e42d1b9c0\": RecentStats: unable to find data in memory cache]" Oct 03 15:00:02 crc kubenswrapper[4959]: I1003 15:00:02.504638 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 03 15:00:02 crc kubenswrapper[4959]: I1003 15:00:02.688515 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.061299 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.181240 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2lpt\" (UniqueName: \"kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt\") pod \"6f4f4059-389d-4e04-90b3-34ec7a502754\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.181652 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume\") pod \"6f4f4059-389d-4e04-90b3-34ec7a502754\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.182012 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume\") pod \"6f4f4059-389d-4e04-90b3-34ec7a502754\" (UID: \"6f4f4059-389d-4e04-90b3-34ec7a502754\") " Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.182575 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume" (OuterVolumeSpecName: "config-volume") pod "6f4f4059-389d-4e04-90b3-34ec7a502754" (UID: "6f4f4059-389d-4e04-90b3-34ec7a502754"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.182898 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6f4f4059-389d-4e04-90b3-34ec7a502754-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.187809 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6f4f4059-389d-4e04-90b3-34ec7a502754" (UID: "6f4f4059-389d-4e04-90b3-34ec7a502754"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.187980 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt" (OuterVolumeSpecName: "kube-api-access-z2lpt") pod "6f4f4059-389d-4e04-90b3-34ec7a502754" (UID: "6f4f4059-389d-4e04-90b3-34ec7a502754"). InnerVolumeSpecName "kube-api-access-z2lpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.284561 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6f4f4059-389d-4e04-90b3-34ec7a502754-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.284611 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2lpt\" (UniqueName: \"kubernetes.io/projected/6f4f4059-389d-4e04-90b3-34ec7a502754-kube-api-access-z2lpt\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.700008 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" path="/var/lib/kubelet/pods/d0caa2dc-a0d2-4354-a2d7-05ea981dac5a/volumes" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.748900 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" event={"ID":"6f4f4059-389d-4e04-90b3-34ec7a502754","Type":"ContainerDied","Data":"d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c"} Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.748944 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d53d67236130bddf9fcf8c922d92021c8bee4ebf65044616e01635ebc993ce0c" Oct 03 15:00:03 crc kubenswrapper[4959]: I1003 15:00:03.748991 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x" Oct 03 15:00:04 crc kubenswrapper[4959]: I1003 15:00:04.198647 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx"] Oct 03 15:00:04 crc kubenswrapper[4959]: I1003 15:00:04.212064 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-7swlx"] Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.192326 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 03 15:00:05 crc kubenswrapper[4959]: E1003 15:00:05.192758 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" containerName="init" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.192778 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" containerName="init" Oct 03 15:00:05 crc kubenswrapper[4959]: E1003 15:00:05.192842 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f4f4059-389d-4e04-90b3-34ec7a502754" containerName="collect-profiles" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.192854 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f4f4059-389d-4e04-90b3-34ec7a502754" containerName="collect-profiles" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.193091 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f4f4059-389d-4e04-90b3-34ec7a502754" containerName="collect-profiles" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.193134 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0caa2dc-a0d2-4354-a2d7-05ea981dac5a" containerName="init" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.194007 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.197167 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.219871 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.321815 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.322177 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb9kp\" (UniqueName: \"kubernetes.io/projected/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-kube-api-access-xb9kp\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.322356 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.424484 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb9kp\" (UniqueName: \"kubernetes.io/projected/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-kube-api-access-xb9kp\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.425312 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.425455 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.428378 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.428417 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/182b6e904521b64262d3d40fe2cf1cf9631a5b536c208126d4437939e49084eb/globalmount\"" pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.433174 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.448971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb9kp\" (UniqueName: \"kubernetes.io/projected/5d211fb6-eef0-4a11-9dcf-9cdd00943b05-kube-api-access-xb9kp\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.459376 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-016cb55b-4e92-4ec7-9250-8d42494490d7\") pod \"ovn-copy-data\" (UID: \"5d211fb6-eef0-4a11-9dcf-9cdd00943b05\") " pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.519561 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 03 15:00:05 crc kubenswrapper[4959]: I1003 15:00:05.699983 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef78006-948e-4e90-ae43-798fd1f38b03" path="/var/lib/kubelet/pods/5ef78006-948e-4e90-ae43-798fd1f38b03/volumes" Oct 03 15:00:06 crc kubenswrapper[4959]: I1003 15:00:06.091148 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 03 15:00:06 crc kubenswrapper[4959]: I1003 15:00:06.773485 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"5d211fb6-eef0-4a11-9dcf-9cdd00943b05","Type":"ContainerStarted","Data":"bebaff56b191cb27de8e4a6c4c9de42921af45c0ba912adf8409c8728409e89a"} Oct 03 15:00:06 crc kubenswrapper[4959]: I1003 15:00:06.773806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"5d211fb6-eef0-4a11-9dcf-9cdd00943b05","Type":"ContainerStarted","Data":"a75ee19089eaf4c8751cb91f05b96ea260c7221456f679c8db49c67cd94f2b01"} Oct 03 15:00:06 crc kubenswrapper[4959]: I1003 15:00:06.802104 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.8020741559999998 podStartE2EDuration="2.802074156s" podCreationTimestamp="2025-10-03 15:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:06.789070508 +0000 UTC m=+5375.992413975" watchObservedRunningTime="2025-10-03 15:00:06.802074156 +0000 UTC m=+5376.005417603" Oct 03 15:00:09 crc kubenswrapper[4959]: I1003 15:00:09.555457 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 15:00:09 crc kubenswrapper[4959]: I1003 15:00:09.637033 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 15:00:09 crc kubenswrapper[4959]: I1003 15:00:09.637363 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="dnsmasq-dns" containerID="cri-o://d7c0b64eb42a181f5939991fa20a25043233055af0fd70d4eb99680609386d46" gracePeriod=10 Oct 03 15:00:09 crc kubenswrapper[4959]: I1003 15:00:09.813423 4959 generic.go:334] "Generic (PLEG): container finished" podID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerID="d7c0b64eb42a181f5939991fa20a25043233055af0fd70d4eb99680609386d46" exitCode=0 Oct 03 15:00:09 crc kubenswrapper[4959]: I1003 15:00:09.813492 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" event={"ID":"1aedf59f-7588-4d1c-ba17-fc60f3c960a6","Type":"ContainerDied","Data":"d7c0b64eb42a181f5939991fa20a25043233055af0fd70d4eb99680609386d46"} Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.097974 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.215923 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config\") pod \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.215980 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc\") pod \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.216184 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrdtq\" (UniqueName: \"kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq\") pod \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\" (UID: \"1aedf59f-7588-4d1c-ba17-fc60f3c960a6\") " Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.221168 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq" (OuterVolumeSpecName: "kube-api-access-wrdtq") pod "1aedf59f-7588-4d1c-ba17-fc60f3c960a6" (UID: "1aedf59f-7588-4d1c-ba17-fc60f3c960a6"). InnerVolumeSpecName "kube-api-access-wrdtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.258326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1aedf59f-7588-4d1c-ba17-fc60f3c960a6" (UID: "1aedf59f-7588-4d1c-ba17-fc60f3c960a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.267487 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config" (OuterVolumeSpecName: "config") pod "1aedf59f-7588-4d1c-ba17-fc60f3c960a6" (UID: "1aedf59f-7588-4d1c-ba17-fc60f3c960a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.317937 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrdtq\" (UniqueName: \"kubernetes.io/projected/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-kube-api-access-wrdtq\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.317967 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.317975 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aedf59f-7588-4d1c-ba17-fc60f3c960a6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.822225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" event={"ID":"1aedf59f-7588-4d1c-ba17-fc60f3c960a6","Type":"ContainerDied","Data":"3e44a15c397cef5227969d50ff1bba148b24337fc57fa533aff2b2a97cdfd823"} Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.822294 4959 scope.go:117] "RemoveContainer" containerID="d7c0b64eb42a181f5939991fa20a25043233055af0fd70d4eb99680609386d46" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.823343 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-964bk" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.863084 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.864579 4959 scope.go:117] "RemoveContainer" containerID="0e6f6dc1a95bcbdfbb0ca3d70c425a54131ef3584e9efd961cf7e05a27eed5ac" Oct 03 15:00:10 crc kubenswrapper[4959]: I1003 15:00:10.873887 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-964bk"] Oct 03 15:00:11 crc kubenswrapper[4959]: I1003 15:00:11.705161 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" path="/var/lib/kubelet/pods/1aedf59f-7588-4d1c-ba17-fc60f3c960a6/volumes" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.231756 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 03 15:00:12 crc kubenswrapper[4959]: E1003 15:00:12.243801 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="dnsmasq-dns" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.243859 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="dnsmasq-dns" Oct 03 15:00:12 crc kubenswrapper[4959]: E1003 15:00:12.243886 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="init" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.243899 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="init" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.244269 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aedf59f-7588-4d1c-ba17-fc60f3c960a6" containerName="dnsmasq-dns" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.246338 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.251020 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9ssvj" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.251308 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.252707 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.260082 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.350909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-config\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.350993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfg6w\" (UniqueName: \"kubernetes.io/projected/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-kube-api-access-cfg6w\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.351015 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.351040 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-scripts\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.351388 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.453473 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.453894 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-config\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.453955 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfg6w\" (UniqueName: \"kubernetes.io/projected/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-kube-api-access-cfg6w\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.454006 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.454029 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-scripts\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.454228 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.455074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-scripts\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.456363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-config\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.464459 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.470662 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfg6w\" (UniqueName: \"kubernetes.io/projected/f7263f47-e4e8-4b64-b043-45d3dfc60c2f-kube-api-access-cfg6w\") pod \"ovn-northd-0\" (UID: \"f7263f47-e4e8-4b64-b043-45d3dfc60c2f\") " pod="openstack/ovn-northd-0" Oct 03 15:00:12 crc kubenswrapper[4959]: I1003 15:00:12.570728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.030255 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.851000 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f7263f47-e4e8-4b64-b043-45d3dfc60c2f","Type":"ContainerStarted","Data":"28baa77d2614b12283667821e8e70ceace5ddf4fe65c77cfd4baa51e9963cedd"} Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.851427 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f7263f47-e4e8-4b64-b043-45d3dfc60c2f","Type":"ContainerStarted","Data":"4c97722e8d92f4c6b95222030060369a7ba44298680cfbc952b28407eb4a1cef"} Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.851460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f7263f47-e4e8-4b64-b043-45d3dfc60c2f","Type":"ContainerStarted","Data":"9d511660f9bd1c02493898566f7f017e43c462ac8ff71b6bf6787d0164dd2f8d"} Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.851523 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 03 15:00:13 crc kubenswrapper[4959]: I1003 15:00:13.886211 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.886169007 podStartE2EDuration="1.886169007s" podCreationTimestamp="2025-10-03 15:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:13.872518943 +0000 UTC m=+5383.075862420" watchObservedRunningTime="2025-10-03 15:00:13.886169007 +0000 UTC m=+5383.089512425" Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.563447 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ql2nc"] Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.564931 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.575367 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ql2nc"] Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.648435 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwgsz\" (UniqueName: \"kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz\") pod \"keystone-db-create-ql2nc\" (UID: \"3f54f365-2baf-43d8-b046-c316d136ea1d\") " pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.750718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwgsz\" (UniqueName: \"kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz\") pod \"keystone-db-create-ql2nc\" (UID: \"3f54f365-2baf-43d8-b046-c316d136ea1d\") " pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.771217 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwgsz\" (UniqueName: \"kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz\") pod \"keystone-db-create-ql2nc\" (UID: \"3f54f365-2baf-43d8-b046-c316d136ea1d\") " pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:17 crc kubenswrapper[4959]: I1003 15:00:17.892430 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:18 crc kubenswrapper[4959]: I1003 15:00:18.320286 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ql2nc"] Oct 03 15:00:18 crc kubenswrapper[4959]: I1003 15:00:18.895807 4959 generic.go:334] "Generic (PLEG): container finished" podID="3f54f365-2baf-43d8-b046-c316d136ea1d" containerID="fc37da1a06d7eeb6bcef1763a3c1a1d2f2b9b23367e175f57fa2b64d5b4f8bf7" exitCode=0 Oct 03 15:00:18 crc kubenswrapper[4959]: I1003 15:00:18.895913 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ql2nc" event={"ID":"3f54f365-2baf-43d8-b046-c316d136ea1d","Type":"ContainerDied","Data":"fc37da1a06d7eeb6bcef1763a3c1a1d2f2b9b23367e175f57fa2b64d5b4f8bf7"} Oct 03 15:00:18 crc kubenswrapper[4959]: I1003 15:00:18.896181 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ql2nc" event={"ID":"3f54f365-2baf-43d8-b046-c316d136ea1d","Type":"ContainerStarted","Data":"e322128e594dcb03f9c92a6ee7de320146d40fd431f95fe7736346fbfb071bd1"} Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.227163 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.291892 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwgsz\" (UniqueName: \"kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz\") pod \"3f54f365-2baf-43d8-b046-c316d136ea1d\" (UID: \"3f54f365-2baf-43d8-b046-c316d136ea1d\") " Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.301477 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz" (OuterVolumeSpecName: "kube-api-access-xwgsz") pod "3f54f365-2baf-43d8-b046-c316d136ea1d" (UID: "3f54f365-2baf-43d8-b046-c316d136ea1d"). InnerVolumeSpecName "kube-api-access-xwgsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.397459 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwgsz\" (UniqueName: \"kubernetes.io/projected/3f54f365-2baf-43d8-b046-c316d136ea1d-kube-api-access-xwgsz\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.928470 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ql2nc" event={"ID":"3f54f365-2baf-43d8-b046-c316d136ea1d","Type":"ContainerDied","Data":"e322128e594dcb03f9c92a6ee7de320146d40fd431f95fe7736346fbfb071bd1"} Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.928526 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ql2nc" Oct 03 15:00:20 crc kubenswrapper[4959]: I1003 15:00:20.928544 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e322128e594dcb03f9c92a6ee7de320146d40fd431f95fe7736346fbfb071bd1" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.162824 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:25 crc kubenswrapper[4959]: E1003 15:00:25.169186 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f54f365-2baf-43d8-b046-c316d136ea1d" containerName="mariadb-database-create" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.169261 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f54f365-2baf-43d8-b046-c316d136ea1d" containerName="mariadb-database-create" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.169624 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f54f365-2baf-43d8-b046-c316d136ea1d" containerName="mariadb-database-create" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.172090 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.172245 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.289403 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.289728 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cjm9\" (UniqueName: \"kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.290160 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.391526 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cjm9\" (UniqueName: \"kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.391906 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.391954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.392476 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.392510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.415688 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cjm9\" (UniqueName: \"kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9\") pod \"certified-operators-ttt29\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:25 crc kubenswrapper[4959]: I1003 15:00:25.511365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:26 crc kubenswrapper[4959]: W1003 15:00:26.044160 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08b9baa5_e80f_405d_89b5_03384258f665.slice/crio-aa3a2811daf3803d07fd13aa3195e4e414854600449efce4d42e25e6febf5157 WatchSource:0}: Error finding container aa3a2811daf3803d07fd13aa3195e4e414854600449efce4d42e25e6febf5157: Status 404 returned error can't find the container with id aa3a2811daf3803d07fd13aa3195e4e414854600449efce4d42e25e6febf5157 Oct 03 15:00:26 crc kubenswrapper[4959]: I1003 15:00:26.049136 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.015796 4959 generic.go:334] "Generic (PLEG): container finished" podID="08b9baa5-e80f-405d-89b5-03384258f665" containerID="0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6" exitCode=0 Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.015921 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerDied","Data":"0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6"} Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.016994 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerStarted","Data":"aa3a2811daf3803d07fd13aa3195e4e414854600449efce4d42e25e6febf5157"} Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.020147 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.551549 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.555535 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.562170 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.631183 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.631395 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjpck\" (UniqueName: \"kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.631430 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.668493 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f61d-account-create-fwfrw"] Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.671871 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.675482 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.676516 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f61d-account-create-fwfrw"] Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.717139 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.733456 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjpck\" (UniqueName: \"kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.733513 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.733607 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccmjx\" (UniqueName: \"kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx\") pod \"keystone-f61d-account-create-fwfrw\" (UID: \"1918b5f3-9683-44e7-ad21-f7cb2935cd7d\") " pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.733639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.737343 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.737895 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.759295 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjpck\" (UniqueName: \"kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck\") pod \"community-operators-fzlnx\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.835382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccmjx\" (UniqueName: \"kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx\") pod \"keystone-f61d-account-create-fwfrw\" (UID: \"1918b5f3-9683-44e7-ad21-f7cb2935cd7d\") " pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.852145 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccmjx\" (UniqueName: \"kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx\") pod \"keystone-f61d-account-create-fwfrw\" (UID: \"1918b5f3-9683-44e7-ad21-f7cb2935cd7d\") " pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.918416 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:27 crc kubenswrapper[4959]: I1003 15:00:27.990285 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:28 crc kubenswrapper[4959]: I1003 15:00:28.026636 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerStarted","Data":"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27"} Oct 03 15:00:28 crc kubenswrapper[4959]: I1003 15:00:28.459292 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:28 crc kubenswrapper[4959]: W1003 15:00:28.464234 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce4a7ec0_4ca2_42fd_b97b_7fc78f82b39c.slice/crio-633eb9ebcfc01a84e7bbdcc24800b28b1a7b2577ff909b91b560a0cf8edda4c0 WatchSource:0}: Error finding container 633eb9ebcfc01a84e7bbdcc24800b28b1a7b2577ff909b91b560a0cf8edda4c0: Status 404 returned error can't find the container with id 633eb9ebcfc01a84e7bbdcc24800b28b1a7b2577ff909b91b560a0cf8edda4c0 Oct 03 15:00:28 crc kubenswrapper[4959]: I1003 15:00:28.500540 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f61d-account-create-fwfrw"] Oct 03 15:00:28 crc kubenswrapper[4959]: W1003 15:00:28.501927 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1918b5f3_9683_44e7_ad21_f7cb2935cd7d.slice/crio-d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d WatchSource:0}: Error finding container d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d: Status 404 returned error can't find the container with id d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.041924 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerID="67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9" exitCode=0 Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.042034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerDied","Data":"67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9"} Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.042446 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerStarted","Data":"633eb9ebcfc01a84e7bbdcc24800b28b1a7b2577ff909b91b560a0cf8edda4c0"} Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.052462 4959 generic.go:334] "Generic (PLEG): container finished" podID="08b9baa5-e80f-405d-89b5-03384258f665" containerID="4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27" exitCode=0 Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.052553 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerDied","Data":"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27"} Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.056382 4959 generic.go:334] "Generic (PLEG): container finished" podID="1918b5f3-9683-44e7-ad21-f7cb2935cd7d" containerID="0955324328fc494d1dd9366aaae06635e38e3be945ae9642faef62cf6ead7456" exitCode=0 Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.056433 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f61d-account-create-fwfrw" event={"ID":"1918b5f3-9683-44e7-ad21-f7cb2935cd7d","Type":"ContainerDied","Data":"0955324328fc494d1dd9366aaae06635e38e3be945ae9642faef62cf6ead7456"} Oct 03 15:00:29 crc kubenswrapper[4959]: I1003 15:00:29.056475 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f61d-account-create-fwfrw" event={"ID":"1918b5f3-9683-44e7-ad21-f7cb2935cd7d","Type":"ContainerStarted","Data":"d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d"} Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.070656 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerStarted","Data":"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095"} Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.074158 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerStarted","Data":"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44"} Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.102740 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ttt29" podStartSLOduration=2.286882401 podStartE2EDuration="5.102716043s" podCreationTimestamp="2025-10-03 15:00:25 +0000 UTC" firstStartedPulling="2025-10-03 15:00:27.019681273 +0000 UTC m=+5396.223024730" lastFinishedPulling="2025-10-03 15:00:29.835514945 +0000 UTC m=+5399.038858372" observedRunningTime="2025-10-03 15:00:30.093655851 +0000 UTC m=+5399.296999318" watchObservedRunningTime="2025-10-03 15:00:30.102716043 +0000 UTC m=+5399.306059480" Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.404732 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.486884 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccmjx\" (UniqueName: \"kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx\") pod \"1918b5f3-9683-44e7-ad21-f7cb2935cd7d\" (UID: \"1918b5f3-9683-44e7-ad21-f7cb2935cd7d\") " Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.499898 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx" (OuterVolumeSpecName: "kube-api-access-ccmjx") pod "1918b5f3-9683-44e7-ad21-f7cb2935cd7d" (UID: "1918b5f3-9683-44e7-ad21-f7cb2935cd7d"). InnerVolumeSpecName "kube-api-access-ccmjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:30 crc kubenswrapper[4959]: I1003 15:00:30.589059 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccmjx\" (UniqueName: \"kubernetes.io/projected/1918b5f3-9683-44e7-ad21-f7cb2935cd7d-kube-api-access-ccmjx\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:31 crc kubenswrapper[4959]: I1003 15:00:31.085532 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f61d-account-create-fwfrw" event={"ID":"1918b5f3-9683-44e7-ad21-f7cb2935cd7d","Type":"ContainerDied","Data":"d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d"} Oct 03 15:00:31 crc kubenswrapper[4959]: I1003 15:00:31.085591 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d604ea88eb6a1209a98097411fd977475f5d584e9470a0b2bd923dd005759e3d" Oct 03 15:00:31 crc kubenswrapper[4959]: I1003 15:00:31.085660 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f61d-account-create-fwfrw" Oct 03 15:00:31 crc kubenswrapper[4959]: I1003 15:00:31.089067 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerID="1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44" exitCode=0 Oct 03 15:00:31 crc kubenswrapper[4959]: I1003 15:00:31.089148 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerDied","Data":"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44"} Oct 03 15:00:32 crc kubenswrapper[4959]: I1003 15:00:32.103529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerStarted","Data":"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4"} Oct 03 15:00:32 crc kubenswrapper[4959]: I1003 15:00:32.135582 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fzlnx" podStartSLOduration=2.651999565 podStartE2EDuration="5.135560735s" podCreationTimestamp="2025-10-03 15:00:27 +0000 UTC" firstStartedPulling="2025-10-03 15:00:29.046802645 +0000 UTC m=+5398.250146102" lastFinishedPulling="2025-10-03 15:00:31.530363855 +0000 UTC m=+5400.733707272" observedRunningTime="2025-10-03 15:00:32.129752192 +0000 UTC m=+5401.333095629" watchObservedRunningTime="2025-10-03 15:00:32.135560735 +0000 UTC m=+5401.338904162" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.132473 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jbwnh"] Oct 03 15:00:33 crc kubenswrapper[4959]: E1003 15:00:33.133118 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1918b5f3-9683-44e7-ad21-f7cb2935cd7d" containerName="mariadb-account-create" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.133134 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1918b5f3-9683-44e7-ad21-f7cb2935cd7d" containerName="mariadb-account-create" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.133319 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1918b5f3-9683-44e7-ad21-f7cb2935cd7d" containerName="mariadb-account-create" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.133956 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.135842 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.137113 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.137248 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.137629 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h2qsb" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.145552 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jbwnh"] Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.232165 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.232459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.232676 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9rnk\" (UniqueName: \"kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.334424 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.334548 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9rnk\" (UniqueName: \"kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.334601 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.340621 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.340788 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.353550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9rnk\" (UniqueName: \"kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk\") pod \"keystone-db-sync-jbwnh\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.500119 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:33 crc kubenswrapper[4959]: I1003 15:00:33.947085 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jbwnh"] Oct 03 15:00:34 crc kubenswrapper[4959]: I1003 15:00:34.119137 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwnh" event={"ID":"54a234cd-5df5-4e16-88fe-78221ec12175","Type":"ContainerStarted","Data":"fa13fde7ca0222dc80df002a31dc026206a42db6517e4ec38ca2f0938e71cfb0"} Oct 03 15:00:35 crc kubenswrapper[4959]: I1003 15:00:35.130690 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwnh" event={"ID":"54a234cd-5df5-4e16-88fe-78221ec12175","Type":"ContainerStarted","Data":"471603a8887b5603e091aeec9862021dc392533e8776b701f1b661441c62771a"} Oct 03 15:00:35 crc kubenswrapper[4959]: I1003 15:00:35.512340 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:35 crc kubenswrapper[4959]: I1003 15:00:35.512420 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:35 crc kubenswrapper[4959]: I1003 15:00:35.583642 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:35 crc kubenswrapper[4959]: I1003 15:00:35.610832 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jbwnh" podStartSLOduration=2.610806191 podStartE2EDuration="2.610806191s" podCreationTimestamp="2025-10-03 15:00:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:35.155433957 +0000 UTC m=+5404.358777374" watchObservedRunningTime="2025-10-03 15:00:35.610806191 +0000 UTC m=+5404.814149648" Oct 03 15:00:36 crc kubenswrapper[4959]: I1003 15:00:36.142036 4959 generic.go:334] "Generic (PLEG): container finished" podID="54a234cd-5df5-4e16-88fe-78221ec12175" containerID="471603a8887b5603e091aeec9862021dc392533e8776b701f1b661441c62771a" exitCode=0 Oct 03 15:00:36 crc kubenswrapper[4959]: I1003 15:00:36.142125 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwnh" event={"ID":"54a234cd-5df5-4e16-88fe-78221ec12175","Type":"ContainerDied","Data":"471603a8887b5603e091aeec9862021dc392533e8776b701f1b661441c62771a"} Oct 03 15:00:36 crc kubenswrapper[4959]: I1003 15:00:36.226628 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:36 crc kubenswrapper[4959]: I1003 15:00:36.284366 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.478496 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.503820 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data\") pod \"54a234cd-5df5-4e16-88fe-78221ec12175\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.503960 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9rnk\" (UniqueName: \"kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk\") pod \"54a234cd-5df5-4e16-88fe-78221ec12175\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.504053 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle\") pod \"54a234cd-5df5-4e16-88fe-78221ec12175\" (UID: \"54a234cd-5df5-4e16-88fe-78221ec12175\") " Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.522138 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk" (OuterVolumeSpecName: "kube-api-access-l9rnk") pod "54a234cd-5df5-4e16-88fe-78221ec12175" (UID: "54a234cd-5df5-4e16-88fe-78221ec12175"). InnerVolumeSpecName "kube-api-access-l9rnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.546409 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54a234cd-5df5-4e16-88fe-78221ec12175" (UID: "54a234cd-5df5-4e16-88fe-78221ec12175"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.559795 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data" (OuterVolumeSpecName: "config-data") pod "54a234cd-5df5-4e16-88fe-78221ec12175" (UID: "54a234cd-5df5-4e16-88fe-78221ec12175"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.605249 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.605286 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54a234cd-5df5-4e16-88fe-78221ec12175-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.605296 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9rnk\" (UniqueName: \"kubernetes.io/projected/54a234cd-5df5-4e16-88fe-78221ec12175-kube-api-access-l9rnk\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.919374 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.919469 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:37 crc kubenswrapper[4959]: I1003 15:00:37.975461 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.168209 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwnh" event={"ID":"54a234cd-5df5-4e16-88fe-78221ec12175","Type":"ContainerDied","Data":"fa13fde7ca0222dc80df002a31dc026206a42db6517e4ec38ca2f0938e71cfb0"} Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.168273 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa13fde7ca0222dc80df002a31dc026206a42db6517e4ec38ca2f0938e71cfb0" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.168446 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwnh" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.169149 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ttt29" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="registry-server" containerID="cri-o://0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095" gracePeriod=2 Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.232916 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.404205 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:00:38 crc kubenswrapper[4959]: E1003 15:00:38.404612 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a234cd-5df5-4e16-88fe-78221ec12175" containerName="keystone-db-sync" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.404627 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a234cd-5df5-4e16-88fe-78221ec12175" containerName="keystone-db-sync" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.404849 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a234cd-5df5-4e16-88fe-78221ec12175" containerName="keystone-db-sync" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.405956 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.435859 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.465690 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hpqml"] Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.466721 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.469468 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h2qsb" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.469505 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.469689 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.469799 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.493864 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hpqml"] Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.534866 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lzk6\" (UniqueName: \"kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.534944 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.534976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.534999 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535074 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535111 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535142 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535173 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535219 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.535280 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lhqx\" (UniqueName: \"kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.541357 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.636981 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lhqx\" (UniqueName: \"kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.637056 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lzk6\" (UniqueName: \"kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.637091 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.637110 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.637126 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.638077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.638135 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.638683 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.638757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.639088 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.639125 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.639149 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.639167 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.640431 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.640504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.645131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.645367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.647345 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.650898 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.654951 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lzk6\" (UniqueName: \"kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.655615 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle\") pod \"keystone-bootstrap-hpqml\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.665160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lhqx\" (UniqueName: \"kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx\") pod \"dnsmasq-dns-58865cd75-fsg97\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.747984 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.765798 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.793802 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.842329 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content\") pod \"08b9baa5-e80f-405d-89b5-03384258f665\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.842472 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities\") pod \"08b9baa5-e80f-405d-89b5-03384258f665\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.842638 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cjm9\" (UniqueName: \"kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9\") pod \"08b9baa5-e80f-405d-89b5-03384258f665\" (UID: \"08b9baa5-e80f-405d-89b5-03384258f665\") " Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.844638 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities" (OuterVolumeSpecName: "utilities") pod "08b9baa5-e80f-405d-89b5-03384258f665" (UID: "08b9baa5-e80f-405d-89b5-03384258f665"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.848890 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9" (OuterVolumeSpecName: "kube-api-access-5cjm9") pod "08b9baa5-e80f-405d-89b5-03384258f665" (UID: "08b9baa5-e80f-405d-89b5-03384258f665"). InnerVolumeSpecName "kube-api-access-5cjm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.944529 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:38 crc kubenswrapper[4959]: I1003 15:00:38.944565 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cjm9\" (UniqueName: \"kubernetes.io/projected/08b9baa5-e80f-405d-89b5-03384258f665-kube-api-access-5cjm9\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.129976 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hpqml"] Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.176063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hpqml" event={"ID":"eb07a5c9-3b77-452c-90d3-98b82aac4666","Type":"ContainerStarted","Data":"1df8231b1993c64bfd46989a837c02e6c19cd2939243cbf67475f7f0e95aa2a0"} Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.178998 4959 generic.go:334] "Generic (PLEG): container finished" podID="08b9baa5-e80f-405d-89b5-03384258f665" containerID="0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095" exitCode=0 Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.179060 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerDied","Data":"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095"} Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.179091 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttt29" event={"ID":"08b9baa5-e80f-405d-89b5-03384258f665","Type":"ContainerDied","Data":"aa3a2811daf3803d07fd13aa3195e4e414854600449efce4d42e25e6febf5157"} Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.179112 4959 scope.go:117] "RemoveContainer" containerID="0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.179117 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttt29" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.208424 4959 scope.go:117] "RemoveContainer" containerID="4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.237834 4959 scope.go:117] "RemoveContainer" containerID="0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.254826 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.264239 4959 scope.go:117] "RemoveContainer" containerID="0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095" Oct 03 15:00:39 crc kubenswrapper[4959]: E1003 15:00:39.264909 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095\": container with ID starting with 0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095 not found: ID does not exist" containerID="0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.264945 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095"} err="failed to get container status \"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095\": rpc error: code = NotFound desc = could not find container \"0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095\": container with ID starting with 0fd082db63d137fcf3aa0dafe877c4d5e31cdf90fb852ba62c68599c7b563095 not found: ID does not exist" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.264967 4959 scope.go:117] "RemoveContainer" containerID="4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27" Oct 03 15:00:39 crc kubenswrapper[4959]: E1003 15:00:39.265356 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27\": container with ID starting with 4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27 not found: ID does not exist" containerID="4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.265386 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27"} err="failed to get container status \"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27\": rpc error: code = NotFound desc = could not find container \"4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27\": container with ID starting with 4b4844c4d56e7275bb19c5f80f3160b96f51712b82995d67b07136c6448a0f27 not found: ID does not exist" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.265402 4959 scope.go:117] "RemoveContainer" containerID="0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6" Oct 03 15:00:39 crc kubenswrapper[4959]: E1003 15:00:39.265637 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6\": container with ID starting with 0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6 not found: ID does not exist" containerID="0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.265662 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6"} err="failed to get container status \"0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6\": rpc error: code = NotFound desc = could not find container \"0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6\": container with ID starting with 0fb088b6a4fe38b38f2d5105e94939d2ba3b0f39c83caa6409ca2f08c92829f6 not found: ID does not exist" Oct 03 15:00:39 crc kubenswrapper[4959]: W1003 15:00:39.269287 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode59377d0_0346_4d14_8b23_2e5d88a8e6ff.slice/crio-c9f89c93092104209a856dda973237ecbf4febf1e67a73c0500f0491a26b739e WatchSource:0}: Error finding container c9f89c93092104209a856dda973237ecbf4febf1e67a73c0500f0491a26b739e: Status 404 returned error can't find the container with id c9f89c93092104209a856dda973237ecbf4febf1e67a73c0500f0491a26b739e Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.561785 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08b9baa5-e80f-405d-89b5-03384258f665" (UID: "08b9baa5-e80f-405d-89b5-03384258f665"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.656986 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b9baa5-e80f-405d-89b5-03384258f665-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.806648 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:39 crc kubenswrapper[4959]: I1003 15:00:39.815362 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ttt29"] Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.189421 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hpqml" event={"ID":"eb07a5c9-3b77-452c-90d3-98b82aac4666","Type":"ContainerStarted","Data":"539f03434b7f8a428d16b72e99611072319094bf66e0c78eea29090d45daa4b8"} Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.191136 4959 generic.go:334] "Generic (PLEG): container finished" podID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerID="cd70d4f50dee87d8663aa8e7587d8d3480641b2fd80243f37604b057716d908a" exitCode=0 Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.191297 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58865cd75-fsg97" event={"ID":"e59377d0-0346-4d14-8b23-2e5d88a8e6ff","Type":"ContainerDied","Data":"cd70d4f50dee87d8663aa8e7587d8d3480641b2fd80243f37604b057716d908a"} Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.191589 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58865cd75-fsg97" event={"ID":"e59377d0-0346-4d14-8b23-2e5d88a8e6ff","Type":"ContainerStarted","Data":"c9f89c93092104209a856dda973237ecbf4febf1e67a73c0500f0491a26b739e"} Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.192921 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fzlnx" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="registry-server" containerID="cri-o://27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4" gracePeriod=2 Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.222801 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hpqml" podStartSLOduration=2.222779761 podStartE2EDuration="2.222779761s" podCreationTimestamp="2025-10-03 15:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:40.220426574 +0000 UTC m=+5409.423770001" watchObservedRunningTime="2025-10-03 15:00:40.222779761 +0000 UTC m=+5409.426123178" Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.620955 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.678867 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities\") pod \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.678996 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjpck\" (UniqueName: \"kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck\") pod \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.679086 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content\") pod \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\" (UID: \"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c\") " Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.681213 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities" (OuterVolumeSpecName: "utilities") pod "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" (UID: "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.686186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck" (OuterVolumeSpecName: "kube-api-access-pjpck") pod "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" (UID: "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c"). InnerVolumeSpecName "kube-api-access-pjpck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.780470 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:40 crc kubenswrapper[4959]: I1003 15:00:40.780504 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjpck\" (UniqueName: \"kubernetes.io/projected/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-kube-api-access-pjpck\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.205548 4959 generic.go:334] "Generic (PLEG): container finished" podID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerID="27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4" exitCode=0 Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.206020 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerDied","Data":"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4"} Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.206061 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fzlnx" event={"ID":"ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c","Type":"ContainerDied","Data":"633eb9ebcfc01a84e7bbdcc24800b28b1a7b2577ff909b91b560a0cf8edda4c0"} Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.206089 4959 scope.go:117] "RemoveContainer" containerID="27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.206815 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fzlnx" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.212992 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58865cd75-fsg97" event={"ID":"e59377d0-0346-4d14-8b23-2e5d88a8e6ff","Type":"ContainerStarted","Data":"7c0e5369c94b466222957921c028bb46715f6cc1b79e946d41f24becd89c5b6f"} Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.213124 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.274712 4959 scope.go:117] "RemoveContainer" containerID="1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.301751 4959 scope.go:117] "RemoveContainer" containerID="67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.361911 4959 scope.go:117] "RemoveContainer" containerID="27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4" Oct 03 15:00:41 crc kubenswrapper[4959]: E1003 15:00:41.362474 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4\": container with ID starting with 27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4 not found: ID does not exist" containerID="27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.362508 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4"} err="failed to get container status \"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4\": rpc error: code = NotFound desc = could not find container \"27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4\": container with ID starting with 27e3333fe94847c70ac96e949777f2550823a51517216e3e0aa338469bc39cb4 not found: ID does not exist" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.362529 4959 scope.go:117] "RemoveContainer" containerID="1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44" Oct 03 15:00:41 crc kubenswrapper[4959]: E1003 15:00:41.362793 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44\": container with ID starting with 1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44 not found: ID does not exist" containerID="1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.362826 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44"} err="failed to get container status \"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44\": rpc error: code = NotFound desc = could not find container \"1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44\": container with ID starting with 1ed5aa0996ba120f42a02d85e005a5774c3c9138f034ff60f0ccbbf39df42e44 not found: ID does not exist" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.362846 4959 scope.go:117] "RemoveContainer" containerID="67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9" Oct 03 15:00:41 crc kubenswrapper[4959]: E1003 15:00:41.363140 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9\": container with ID starting with 67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9 not found: ID does not exist" containerID="67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.363167 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9"} err="failed to get container status \"67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9\": rpc error: code = NotFound desc = could not find container \"67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9\": container with ID starting with 67b537c801af3bdafbbb4f6fbb57d8317c3347e2ce8dfa7bae3d95c196e2fed9 not found: ID does not exist" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.692519 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" (UID: "ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.699734 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.706181 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b9baa5-e80f-405d-89b5-03384258f665" path="/var/lib/kubelet/pods/08b9baa5-e80f-405d-89b5-03384258f665/volumes" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.832088 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58865cd75-fsg97" podStartSLOduration=3.832067329 podStartE2EDuration="3.832067329s" podCreationTimestamp="2025-10-03 15:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:41.241279993 +0000 UTC m=+5410.444623470" watchObservedRunningTime="2025-10-03 15:00:41.832067329 +0000 UTC m=+5411.035410746" Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.833452 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:41 crc kubenswrapper[4959]: I1003 15:00:41.840728 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fzlnx"] Oct 03 15:00:43 crc kubenswrapper[4959]: I1003 15:00:43.238546 4959 generic.go:334] "Generic (PLEG): container finished" podID="eb07a5c9-3b77-452c-90d3-98b82aac4666" containerID="539f03434b7f8a428d16b72e99611072319094bf66e0c78eea29090d45daa4b8" exitCode=0 Oct 03 15:00:43 crc kubenswrapper[4959]: I1003 15:00:43.238650 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hpqml" event={"ID":"eb07a5c9-3b77-452c-90d3-98b82aac4666","Type":"ContainerDied","Data":"539f03434b7f8a428d16b72e99611072319094bf66e0c78eea29090d45daa4b8"} Oct 03 15:00:43 crc kubenswrapper[4959]: I1003 15:00:43.715788 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" path="/var/lib/kubelet/pods/ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c/volumes" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.620996 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751546 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lzk6\" (UniqueName: \"kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751634 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751694 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751737 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751778 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.751806 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle\") pod \"eb07a5c9-3b77-452c-90d3-98b82aac4666\" (UID: \"eb07a5c9-3b77-452c-90d3-98b82aac4666\") " Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.758178 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts" (OuterVolumeSpecName: "scripts") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.758499 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.759182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6" (OuterVolumeSpecName: "kube-api-access-7lzk6") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "kube-api-access-7lzk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.760751 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.774879 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data" (OuterVolumeSpecName: "config-data") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.780025 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb07a5c9-3b77-452c-90d3-98b82aac4666" (UID: "eb07a5c9-3b77-452c-90d3-98b82aac4666"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855103 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855145 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855162 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855177 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855211 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lzk6\" (UniqueName: \"kubernetes.io/projected/eb07a5c9-3b77-452c-90d3-98b82aac4666-kube-api-access-7lzk6\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:44 crc kubenswrapper[4959]: I1003 15:00:44.855227 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb07a5c9-3b77-452c-90d3-98b82aac4666-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.262574 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hpqml" event={"ID":"eb07a5c9-3b77-452c-90d3-98b82aac4666","Type":"ContainerDied","Data":"1df8231b1993c64bfd46989a837c02e6c19cd2939243cbf67475f7f0e95aa2a0"} Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.262614 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hpqml" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.262635 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df8231b1993c64bfd46989a837c02e6c19cd2939243cbf67475f7f0e95aa2a0" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.431596 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hpqml"] Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.440139 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hpqml"] Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517311 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5vgl4"] Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517699 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="extract-utilities" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517723 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="extract-utilities" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517741 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb07a5c9-3b77-452c-90d3-98b82aac4666" containerName="keystone-bootstrap" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517749 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb07a5c9-3b77-452c-90d3-98b82aac4666" containerName="keystone-bootstrap" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517768 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="extract-content" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517775 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="extract-content" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517793 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="extract-content" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517800 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="extract-content" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517817 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517824 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517833 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="extract-utilities" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517842 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="extract-utilities" Oct 03 15:00:45 crc kubenswrapper[4959]: E1003 15:00:45.517853 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.517861 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.518039 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb07a5c9-3b77-452c-90d3-98b82aac4666" containerName="keystone-bootstrap" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.518066 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b9baa5-e80f-405d-89b5-03384258f665" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.518079 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4a7ec0-4ca2-42fd-b97b-7fc78f82b39c" containerName="registry-server" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.518743 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.521806 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.522225 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.522410 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.522754 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h2qsb" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.531116 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5vgl4"] Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.565713 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.566057 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.566140 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgch\" (UniqueName: \"kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.566308 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.566354 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.566404 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668520 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgch\" (UniqueName: \"kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668603 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668626 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668647 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.668698 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.674572 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.674712 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.675061 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.675371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.677412 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.687559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgch\" (UniqueName: \"kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch\") pod \"keystone-bootstrap-5vgl4\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.697716 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb07a5c9-3b77-452c-90d3-98b82aac4666" path="/var/lib/kubelet/pods/eb07a5c9-3b77-452c-90d3-98b82aac4666/volumes" Oct 03 15:00:45 crc kubenswrapper[4959]: I1003 15:00:45.843948 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:46 crc kubenswrapper[4959]: I1003 15:00:46.393139 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5vgl4"] Oct 03 15:00:46 crc kubenswrapper[4959]: W1003 15:00:46.403992 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc398778d_bdd5_49c1_a1b8_348cd53cf86a.slice/crio-807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c WatchSource:0}: Error finding container 807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c: Status 404 returned error can't find the container with id 807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c Oct 03 15:00:47 crc kubenswrapper[4959]: I1003 15:00:47.286861 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5vgl4" event={"ID":"c398778d-bdd5-49c1-a1b8-348cd53cf86a","Type":"ContainerStarted","Data":"134d4662d6c6e20af268d24f25cb84fcd5ad640260002e4a4fc2c65c92e99e85"} Oct 03 15:00:47 crc kubenswrapper[4959]: I1003 15:00:47.287293 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5vgl4" event={"ID":"c398778d-bdd5-49c1-a1b8-348cd53cf86a","Type":"ContainerStarted","Data":"807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c"} Oct 03 15:00:47 crc kubenswrapper[4959]: I1003 15:00:47.327845 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5vgl4" podStartSLOduration=2.327820144 podStartE2EDuration="2.327820144s" podCreationTimestamp="2025-10-03 15:00:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:47.321518731 +0000 UTC m=+5416.524862188" watchObservedRunningTime="2025-10-03 15:00:47.327820144 +0000 UTC m=+5416.531163591" Oct 03 15:00:48 crc kubenswrapper[4959]: I1003 15:00:48.767485 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:00:48 crc kubenswrapper[4959]: I1003 15:00:48.837772 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 15:00:48 crc kubenswrapper[4959]: I1003 15:00:48.838012 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="dnsmasq-dns" containerID="cri-o://8625b7a263056a2ed3fbbe2864e024162970a01f5177078d576ee2ac2f28b288" gracePeriod=10 Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.304649 4959 generic.go:334] "Generic (PLEG): container finished" podID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerID="8625b7a263056a2ed3fbbe2864e024162970a01f5177078d576ee2ac2f28b288" exitCode=0 Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.304953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" event={"ID":"a39a3471-a59a-43cf-9e3c-db53b8c0c63e","Type":"ContainerDied","Data":"8625b7a263056a2ed3fbbe2864e024162970a01f5177078d576ee2ac2f28b288"} Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.553945 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.12:5353: connect: connection refused" Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.828853 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.961810 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc\") pod \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.961960 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb\") pod \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.962058 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpj94\" (UniqueName: \"kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94\") pod \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.962077 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config\") pod \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.962093 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb\") pod \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\" (UID: \"a39a3471-a59a-43cf-9e3c-db53b8c0c63e\") " Oct 03 15:00:49 crc kubenswrapper[4959]: I1003 15:00:49.970390 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94" (OuterVolumeSpecName: "kube-api-access-jpj94") pod "a39a3471-a59a-43cf-9e3c-db53b8c0c63e" (UID: "a39a3471-a59a-43cf-9e3c-db53b8c0c63e"). InnerVolumeSpecName "kube-api-access-jpj94". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.011125 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config" (OuterVolumeSpecName: "config") pod "a39a3471-a59a-43cf-9e3c-db53b8c0c63e" (UID: "a39a3471-a59a-43cf-9e3c-db53b8c0c63e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.014261 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a39a3471-a59a-43cf-9e3c-db53b8c0c63e" (UID: "a39a3471-a59a-43cf-9e3c-db53b8c0c63e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.032179 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a39a3471-a59a-43cf-9e3c-db53b8c0c63e" (UID: "a39a3471-a59a-43cf-9e3c-db53b8c0c63e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.037391 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a39a3471-a59a-43cf-9e3c-db53b8c0c63e" (UID: "a39a3471-a59a-43cf-9e3c-db53b8c0c63e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.063514 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.063573 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpj94\" (UniqueName: \"kubernetes.io/projected/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-kube-api-access-jpj94\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.063584 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.063648 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.063660 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a39a3471-a59a-43cf-9e3c-db53b8c0c63e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.314482 4959 generic.go:334] "Generic (PLEG): container finished" podID="c398778d-bdd5-49c1-a1b8-348cd53cf86a" containerID="134d4662d6c6e20af268d24f25cb84fcd5ad640260002e4a4fc2c65c92e99e85" exitCode=0 Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.314584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5vgl4" event={"ID":"c398778d-bdd5-49c1-a1b8-348cd53cf86a","Type":"ContainerDied","Data":"134d4662d6c6e20af268d24f25cb84fcd5ad640260002e4a4fc2c65c92e99e85"} Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.317607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" event={"ID":"a39a3471-a59a-43cf-9e3c-db53b8c0c63e","Type":"ContainerDied","Data":"567b01a34906bfc93d97e03bb9219cbea9aa01e2c2c7318a5e5cccbac5e1807f"} Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.317643 4959 scope.go:117] "RemoveContainer" containerID="8625b7a263056a2ed3fbbe2864e024162970a01f5177078d576ee2ac2f28b288" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.317802 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864bc46885-gvsdp" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.360717 4959 scope.go:117] "RemoveContainer" containerID="d3b1f51f3a25001eeb9ef50f61c5abf3df8647ec6e954966c59a0f8cf6ac8af7" Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.394027 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 15:00:50 crc kubenswrapper[4959]: I1003 15:00:50.404701 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864bc46885-gvsdp"] Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.668575 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.720173 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" path="/var/lib/kubelet/pods/a39a3471-a59a-43cf-9e3c-db53b8c0c63e/volumes" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795647 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795685 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795758 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgch\" (UniqueName: \"kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795821 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795900 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.795953 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data\") pod \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\" (UID: \"c398778d-bdd5-49c1-a1b8-348cd53cf86a\") " Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.801068 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.801701 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch" (OuterVolumeSpecName: "kube-api-access-rrgch") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "kube-api-access-rrgch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.801711 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts" (OuterVolumeSpecName: "scripts") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.802502 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.817599 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data" (OuterVolumeSpecName: "config-data") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.841547 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c398778d-bdd5-49c1-a1b8-348cd53cf86a" (UID: "c398778d-bdd5-49c1-a1b8-348cd53cf86a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897211 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgch\" (UniqueName: \"kubernetes.io/projected/c398778d-bdd5-49c1-a1b8-348cd53cf86a-kube-api-access-rrgch\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897246 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897259 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897274 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897284 4959 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:51 crc kubenswrapper[4959]: I1003 15:00:51.897319 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c398778d-bdd5-49c1-a1b8-348cd53cf86a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.349597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5vgl4" event={"ID":"c398778d-bdd5-49c1-a1b8-348cd53cf86a","Type":"ContainerDied","Data":"807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c"} Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.349654 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807f3c4a1d23b62f01d5bac8a78a85263f57d89d8e95530509d11b6b5770a04c" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.349736 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5vgl4" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.461279 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8fdb48fc-6jdsh"] Oct 03 15:00:52 crc kubenswrapper[4959]: E1003 15:00:52.461810 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="init" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.461840 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="init" Oct 03 15:00:52 crc kubenswrapper[4959]: E1003 15:00:52.461870 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c398778d-bdd5-49c1-a1b8-348cd53cf86a" containerName="keystone-bootstrap" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.461879 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c398778d-bdd5-49c1-a1b8-348cd53cf86a" containerName="keystone-bootstrap" Oct 03 15:00:52 crc kubenswrapper[4959]: E1003 15:00:52.461916 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="dnsmasq-dns" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.461927 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="dnsmasq-dns" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.462128 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a39a3471-a59a-43cf-9e3c-db53b8c0c63e" containerName="dnsmasq-dns" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.462163 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c398778d-bdd5-49c1-a1b8-348cd53cf86a" containerName="keystone-bootstrap" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.463060 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.466174 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-h2qsb" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.466700 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.467019 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.473372 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.481548 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8fdb48fc-6jdsh"] Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.510600 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-config-data\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.510680 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-credential-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.512096 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-combined-ca-bundle\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.512161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-fernet-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.512236 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-scripts\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.512260 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5q2f\" (UniqueName: \"kubernetes.io/projected/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-kube-api-access-d5q2f\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.613858 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-credential-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.613921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-combined-ca-bundle\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.613962 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-fernet-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.614007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-scripts\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.614049 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5q2f\" (UniqueName: \"kubernetes.io/projected/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-kube-api-access-d5q2f\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.614128 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-config-data\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.618839 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-credential-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.619321 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-combined-ca-bundle\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.627175 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-config-data\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.627794 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-fernet-keys\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.637294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5q2f\" (UniqueName: \"kubernetes.io/projected/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-kube-api-access-d5q2f\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.637445 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f-scripts\") pod \"keystone-8fdb48fc-6jdsh\" (UID: \"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f\") " pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:52 crc kubenswrapper[4959]: I1003 15:00:52.805711 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:53 crc kubenswrapper[4959]: I1003 15:00:53.354061 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8fdb48fc-6jdsh"] Oct 03 15:00:54 crc kubenswrapper[4959]: I1003 15:00:54.370132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdb48fc-6jdsh" event={"ID":"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f","Type":"ContainerStarted","Data":"ee2e2bef389033c0dd683a5778cb9c477070e8a9f2e1c49559bdeb5209d0d388"} Oct 03 15:00:54 crc kubenswrapper[4959]: I1003 15:00:54.370715 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:00:54 crc kubenswrapper[4959]: I1003 15:00:54.370732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8fdb48fc-6jdsh" event={"ID":"47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f","Type":"ContainerStarted","Data":"0d28752e1921ae705d6fbb18bb3ce211fb1b08684ca05bcefd2218be14d45584"} Oct 03 15:00:54 crc kubenswrapper[4959]: I1003 15:00:54.400888 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8fdb48fc-6jdsh" podStartSLOduration=2.400863305 podStartE2EDuration="2.400863305s" podCreationTimestamp="2025-10-03 15:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:00:54.39245474 +0000 UTC m=+5423.595798157" watchObservedRunningTime="2025-10-03 15:00:54.400863305 +0000 UTC m=+5423.604206732" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.157028 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325061-jxq5l"] Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.158681 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.188642 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325061-jxq5l"] Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.257779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6vl\" (UniqueName: \"kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.258115 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.258224 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.258327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.359284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.359343 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.359368 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r6vl\" (UniqueName: \"kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.359507 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.366953 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.367103 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.368276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.383519 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r6vl\" (UniqueName: \"kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl\") pod \"keystone-cron-29325061-jxq5l\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.488685 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:00 crc kubenswrapper[4959]: I1003 15:01:00.930882 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325061-jxq5l"] Oct 03 15:01:01 crc kubenswrapper[4959]: I1003 15:01:01.439000 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325061-jxq5l" event={"ID":"7403bf5f-ba18-42cb-8d44-f5ed5d463efd","Type":"ContainerStarted","Data":"a78d6fda6ae90d52f31574eb67b3eabccd19f90032af7ae3673ced82d6a8c6ab"} Oct 03 15:01:01 crc kubenswrapper[4959]: I1003 15:01:01.439368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325061-jxq5l" event={"ID":"7403bf5f-ba18-42cb-8d44-f5ed5d463efd","Type":"ContainerStarted","Data":"fba6c909c322937e9372ba09d766d2abf178a1052ccb5ed439751ba1346f14aa"} Oct 03 15:01:01 crc kubenswrapper[4959]: I1003 15:01:01.462950 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325061-jxq5l" podStartSLOduration=1.462930537 podStartE2EDuration="1.462930537s" podCreationTimestamp="2025-10-03 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:01:01.456597532 +0000 UTC m=+5430.659940979" watchObservedRunningTime="2025-10-03 15:01:01.462930537 +0000 UTC m=+5430.666273954" Oct 03 15:01:03 crc kubenswrapper[4959]: I1003 15:01:03.460864 4959 generic.go:334] "Generic (PLEG): container finished" podID="7403bf5f-ba18-42cb-8d44-f5ed5d463efd" containerID="a78d6fda6ae90d52f31574eb67b3eabccd19f90032af7ae3673ced82d6a8c6ab" exitCode=0 Oct 03 15:01:03 crc kubenswrapper[4959]: I1003 15:01:03.460958 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325061-jxq5l" event={"ID":"7403bf5f-ba18-42cb-8d44-f5ed5d463efd","Type":"ContainerDied","Data":"a78d6fda6ae90d52f31574eb67b3eabccd19f90032af7ae3673ced82d6a8c6ab"} Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.841777 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.961134 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys\") pod \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.961265 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r6vl\" (UniqueName: \"kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl\") pod \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.961448 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data\") pod \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.961557 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle\") pod \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\" (UID: \"7403bf5f-ba18-42cb-8d44-f5ed5d463efd\") " Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.970632 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl" (OuterVolumeSpecName: "kube-api-access-6r6vl") pod "7403bf5f-ba18-42cb-8d44-f5ed5d463efd" (UID: "7403bf5f-ba18-42cb-8d44-f5ed5d463efd"). InnerVolumeSpecName "kube-api-access-6r6vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.971532 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7403bf5f-ba18-42cb-8d44-f5ed5d463efd" (UID: "7403bf5f-ba18-42cb-8d44-f5ed5d463efd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:01:04 crc kubenswrapper[4959]: I1003 15:01:04.989756 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7403bf5f-ba18-42cb-8d44-f5ed5d463efd" (UID: "7403bf5f-ba18-42cb-8d44-f5ed5d463efd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.041979 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data" (OuterVolumeSpecName: "config-data") pod "7403bf5f-ba18-42cb-8d44-f5ed5d463efd" (UID: "7403bf5f-ba18-42cb-8d44-f5ed5d463efd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.063381 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.063404 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r6vl\" (UniqueName: \"kubernetes.io/projected/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-kube-api-access-6r6vl\") on node \"crc\" DevicePath \"\"" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.063414 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.063421 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7403bf5f-ba18-42cb-8d44-f5ed5d463efd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.258799 4959 scope.go:117] "RemoveContainer" containerID="591346f15def5fc51a579ec324f7321064125f246861471bb162fdffe85b79b2" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.288303 4959 scope.go:117] "RemoveContainer" containerID="37d2db0004bd25d622ca66d7bedcae1b058d8879f4ea407e1188921432b60a3e" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.307903 4959 scope.go:117] "RemoveContainer" containerID="72959b2d8bc8e0fb5b2ae1654f4cc40d682e221747c87180c11a01f00c8aab1f" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.338373 4959 scope.go:117] "RemoveContainer" containerID="aaab30786b377eecb4418fc42d5429a04cbd04b9ff4646fcce80433647bde683" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.364279 4959 scope.go:117] "RemoveContainer" containerID="6e1623d214c36de6c2178ff445457c448cc21feacc3f66e6e1a893c43e1db9d6" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.385866 4959 scope.go:117] "RemoveContainer" containerID="9b1a0cb0860a381daec0aa0dd4e4bc1f2c42598e3e0916acd4022f2553168334" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.408979 4959 scope.go:117] "RemoveContainer" containerID="95aeaa7374f1b9d3ac416e7141ffce96300d1bcdac5bed3265593e5e06f694a3" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.432785 4959 scope.go:117] "RemoveContainer" containerID="c27b4b059704e1d5eafa57548fb8ce9134fb5032cd5ddc79d8a73b05c87c88ee" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.458081 4959 scope.go:117] "RemoveContainer" containerID="b4dcf4da2beface3decaa5295dd894e223f459d934ac2db231f3a284089363bd" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.528160 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325061-jxq5l" event={"ID":"7403bf5f-ba18-42cb-8d44-f5ed5d463efd","Type":"ContainerDied","Data":"fba6c909c322937e9372ba09d766d2abf178a1052ccb5ed439751ba1346f14aa"} Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.528231 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba6c909c322937e9372ba09d766d2abf178a1052ccb5ed439751ba1346f14aa" Oct 03 15:01:05 crc kubenswrapper[4959]: I1003 15:01:05.528235 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325061-jxq5l" Oct 03 15:01:06 crc kubenswrapper[4959]: I1003 15:01:06.044209 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:01:06 crc kubenswrapper[4959]: I1003 15:01:06.044273 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:01:24 crc kubenswrapper[4959]: I1003 15:01:24.161116 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-8fdb48fc-6jdsh" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.720242 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: E1003 15:01:27.722042 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7403bf5f-ba18-42cb-8d44-f5ed5d463efd" containerName="keystone-cron" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.722122 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7403bf5f-ba18-42cb-8d44-f5ed5d463efd" containerName="keystone-cron" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.722374 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7403bf5f-ba18-42cb-8d44-f5ed5d463efd" containerName="keystone-cron" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.723062 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.725583 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rxktg" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.725852 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.726081 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.732478 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.737624 4959 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e76f43d8-067c-4b09-b4ac-051d6247c21a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T15:01:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T15:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T15:01:27Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T15:01:27Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bs26b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T15:01:27Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.742676 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.750581 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: E1003 15:01:27.760647 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-bs26b openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[kube-api-access-bs26b openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="e76f43d8-067c-4b09-b4ac-051d6247c21a" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.774651 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.775649 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.793731 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.799083 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e76f43d8-067c-4b09-b4ac-051d6247c21a" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.857599 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.857839 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.857965 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjwp9\" (UniqueName: \"kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.960021 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.960119 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjwp9\" (UniqueName: \"kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.960599 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.961799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.966510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:27 crc kubenswrapper[4959]: I1003 15:01:27.976685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjwp9\" (UniqueName: \"kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9\") pod \"openstackclient\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " pod="openstack/openstackclient" Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.095815 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.520433 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.772157 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42","Type":"ContainerStarted","Data":"c7ea5b66e58c444ff1381a4fcf5bfc87affa757002b29fff1d266e600f74daaa"} Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.772499 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42","Type":"ContainerStarted","Data":"ee9faced55b8c63f6b93ee49611d208fb8d23f9e9ee261859ccadbf8b55cd309"} Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.772179 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.794927 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.799643 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e76f43d8-067c-4b09-b4ac-051d6247c21a" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" Oct 03 15:01:28 crc kubenswrapper[4959]: I1003 15:01:28.800753 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.80073077 podStartE2EDuration="1.80073077s" podCreationTimestamp="2025-10-03 15:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:01:28.79379987 +0000 UTC m=+5457.997143297" watchObservedRunningTime="2025-10-03 15:01:28.80073077 +0000 UTC m=+5458.004074187" Oct 03 15:01:29 crc kubenswrapper[4959]: I1003 15:01:29.697465 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76f43d8-067c-4b09-b4ac-051d6247c21a" path="/var/lib/kubelet/pods/e76f43d8-067c-4b09-b4ac-051d6247c21a/volumes" Oct 03 15:01:29 crc kubenswrapper[4959]: I1003 15:01:29.781442 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:01:29 crc kubenswrapper[4959]: I1003 15:01:29.787935 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e76f43d8-067c-4b09-b4ac-051d6247c21a" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" Oct 03 15:01:36 crc kubenswrapper[4959]: I1003 15:01:36.044529 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:01:36 crc kubenswrapper[4959]: I1003 15:01:36.044973 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:02:06 crc kubenswrapper[4959]: I1003 15:02:06.044885 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:02:06 crc kubenswrapper[4959]: I1003 15:02:06.045527 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:02:06 crc kubenswrapper[4959]: I1003 15:02:06.045584 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:02:06 crc kubenswrapper[4959]: I1003 15:02:06.046600 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:02:06 crc kubenswrapper[4959]: I1003 15:02:06.046691 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e" gracePeriod=600 Oct 03 15:02:07 crc kubenswrapper[4959]: I1003 15:02:07.114026 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e" exitCode=0 Oct 03 15:02:07 crc kubenswrapper[4959]: I1003 15:02:07.114108 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e"} Oct 03 15:02:07 crc kubenswrapper[4959]: I1003 15:02:07.114378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505"} Oct 03 15:02:07 crc kubenswrapper[4959]: I1003 15:02:07.114398 4959 scope.go:117] "RemoveContainer" containerID="e2eb98e003f27d5132329430337fb3bd12d18b50f827f825926797b0381fba8f" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.467722 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.470363 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.477968 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.625297 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.625622 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tdr\" (UniqueName: \"kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.625648 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.727229 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.727305 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tdr\" (UniqueName: \"kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.727333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.727889 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.727912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.753048 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tdr\" (UniqueName: \"kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr\") pod \"redhat-operators-kglwm\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:41 crc kubenswrapper[4959]: I1003 15:02:41.790403 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:42 crc kubenswrapper[4959]: I1003 15:02:42.246173 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:42 crc kubenswrapper[4959]: I1003 15:02:42.450846 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerStarted","Data":"395e2c97afa66d151dd95aa125edca00be5bc83194b87e0b84f0c91e61499b83"} Oct 03 15:02:43 crc kubenswrapper[4959]: I1003 15:02:43.463900 4959 generic.go:334] "Generic (PLEG): container finished" podID="597deac7-221b-4cd6-88fe-d7d628790248" containerID="64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b" exitCode=0 Oct 03 15:02:43 crc kubenswrapper[4959]: I1003 15:02:43.463991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerDied","Data":"64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b"} Oct 03 15:02:45 crc kubenswrapper[4959]: I1003 15:02:45.487059 4959 generic.go:334] "Generic (PLEG): container finished" podID="597deac7-221b-4cd6-88fe-d7d628790248" containerID="3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c" exitCode=0 Oct 03 15:02:45 crc kubenswrapper[4959]: I1003 15:02:45.487162 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerDied","Data":"3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c"} Oct 03 15:02:46 crc kubenswrapper[4959]: I1003 15:02:46.499304 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerStarted","Data":"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8"} Oct 03 15:02:46 crc kubenswrapper[4959]: I1003 15:02:46.524920 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kglwm" podStartSLOduration=2.9939935269999998 podStartE2EDuration="5.524900466s" podCreationTimestamp="2025-10-03 15:02:41 +0000 UTC" firstStartedPulling="2025-10-03 15:02:43.466590823 +0000 UTC m=+5532.669934290" lastFinishedPulling="2025-10-03 15:02:45.997497782 +0000 UTC m=+5535.200841229" observedRunningTime="2025-10-03 15:02:46.521227197 +0000 UTC m=+5535.724570634" watchObservedRunningTime="2025-10-03 15:02:46.524900466 +0000 UTC m=+5535.728243883" Oct 03 15:02:51 crc kubenswrapper[4959]: I1003 15:02:51.790710 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:51 crc kubenswrapper[4959]: I1003 15:02:51.791304 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:51 crc kubenswrapper[4959]: I1003 15:02:51.868697 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:52 crc kubenswrapper[4959]: I1003 15:02:52.610246 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:52 crc kubenswrapper[4959]: I1003 15:02:52.666257 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:54 crc kubenswrapper[4959]: I1003 15:02:54.578131 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kglwm" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="registry-server" containerID="cri-o://02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8" gracePeriod=2 Oct 03 15:02:54 crc kubenswrapper[4959]: I1003 15:02:54.982783 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:02:54 crc kubenswrapper[4959]: I1003 15:02:54.989414 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.008784 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.051003 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.072403 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content\") pod \"597deac7-221b-4cd6-88fe-d7d628790248\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.073273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities\") pod \"597deac7-221b-4cd6-88fe-d7d628790248\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.073488 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8tdr\" (UniqueName: \"kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr\") pod \"597deac7-221b-4cd6-88fe-d7d628790248\" (UID: \"597deac7-221b-4cd6-88fe-d7d628790248\") " Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.073884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.074020 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.074620 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities" (OuterVolumeSpecName: "utilities") pod "597deac7-221b-4cd6-88fe-d7d628790248" (UID: "597deac7-221b-4cd6-88fe-d7d628790248"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.075376 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28kl5\" (UniqueName: \"kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.075784 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.080382 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr" (OuterVolumeSpecName: "kube-api-access-n8tdr") pod "597deac7-221b-4cd6-88fe-d7d628790248" (UID: "597deac7-221b-4cd6-88fe-d7d628790248"). InnerVolumeSpecName "kube-api-access-n8tdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.176009 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "597deac7-221b-4cd6-88fe-d7d628790248" (UID: "597deac7-221b-4cd6-88fe-d7d628790248"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.177718 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.177781 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.177844 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28kl5\" (UniqueName: \"kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.177941 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/597deac7-221b-4cd6-88fe-d7d628790248-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.177956 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8tdr\" (UniqueName: \"kubernetes.io/projected/597deac7-221b-4cd6-88fe-d7d628790248-kube-api-access-n8tdr\") on node \"crc\" DevicePath \"\"" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.178268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.178371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.196907 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28kl5\" (UniqueName: \"kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5\") pod \"redhat-marketplace-fwbt7\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.363875 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.591002 4959 generic.go:334] "Generic (PLEG): container finished" podID="597deac7-221b-4cd6-88fe-d7d628790248" containerID="02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8" exitCode=0 Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.591342 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerDied","Data":"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8"} Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.591384 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kglwm" event={"ID":"597deac7-221b-4cd6-88fe-d7d628790248","Type":"ContainerDied","Data":"395e2c97afa66d151dd95aa125edca00be5bc83194b87e0b84f0c91e61499b83"} Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.591411 4959 scope.go:117] "RemoveContainer" containerID="02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.591598 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kglwm" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.618556 4959 scope.go:117] "RemoveContainer" containerID="3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.633461 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.645909 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kglwm"] Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.657539 4959 scope.go:117] "RemoveContainer" containerID="64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.688428 4959 scope.go:117] "RemoveContainer" containerID="02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8" Oct 03 15:02:55 crc kubenswrapper[4959]: E1003 15:02:55.689322 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8\": container with ID starting with 02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8 not found: ID does not exist" containerID="02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.689358 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8"} err="failed to get container status \"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8\": rpc error: code = NotFound desc = could not find container \"02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8\": container with ID starting with 02912239a1085b8335bd51e7daaf42845b4541f57a35a0ac5e919009c1599ad8 not found: ID does not exist" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.689386 4959 scope.go:117] "RemoveContainer" containerID="3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c" Oct 03 15:02:55 crc kubenswrapper[4959]: E1003 15:02:55.689988 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c\": container with ID starting with 3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c not found: ID does not exist" containerID="3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.690007 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c"} err="failed to get container status \"3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c\": rpc error: code = NotFound desc = could not find container \"3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c\": container with ID starting with 3b0afa3a4b7bd0131fe0d58809384e92d9a4f5d113253b06cf0827562802cd6c not found: ID does not exist" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.690019 4959 scope.go:117] "RemoveContainer" containerID="64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.696071 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="597deac7-221b-4cd6-88fe-d7d628790248" path="/var/lib/kubelet/pods/597deac7-221b-4cd6-88fe-d7d628790248/volumes" Oct 03 15:02:55 crc kubenswrapper[4959]: E1003 15:02:55.696895 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b\": container with ID starting with 64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b not found: ID does not exist" containerID="64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.696917 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b"} err="failed to get container status \"64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b\": rpc error: code = NotFound desc = could not find container \"64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b\": container with ID starting with 64ddec9204ed58835891b9f7535985fb01af2c8cf8e36f3239501d4b86c0f37b not found: ID does not exist" Oct 03 15:02:55 crc kubenswrapper[4959]: E1003 15:02:55.823248 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod597deac7_221b_4cd6_88fe_d7d628790248.slice/crio-395e2c97afa66d151dd95aa125edca00be5bc83194b87e0b84f0c91e61499b83\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod597deac7_221b_4cd6_88fe_d7d628790248.slice\": RecentStats: unable to find data in memory cache]" Oct 03 15:02:55 crc kubenswrapper[4959]: I1003 15:02:55.840552 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:02:55 crc kubenswrapper[4959]: W1003 15:02:55.845258 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda205c470_f37a_4a98_92a6_0dca74424d9d.slice/crio-c46034cfda9b95f8c89faa595fdd5be90ae3cd7f85156e830c1012b7c7682508 WatchSource:0}: Error finding container c46034cfda9b95f8c89faa595fdd5be90ae3cd7f85156e830c1012b7c7682508: Status 404 returned error can't find the container with id c46034cfda9b95f8c89faa595fdd5be90ae3cd7f85156e830c1012b7c7682508 Oct 03 15:02:56 crc kubenswrapper[4959]: I1003 15:02:56.601676 4959 generic.go:334] "Generic (PLEG): container finished" podID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerID="3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b" exitCode=0 Oct 03 15:02:56 crc kubenswrapper[4959]: I1003 15:02:56.601727 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerDied","Data":"3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b"} Oct 03 15:02:56 crc kubenswrapper[4959]: I1003 15:02:56.602047 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerStarted","Data":"c46034cfda9b95f8c89faa595fdd5be90ae3cd7f85156e830c1012b7c7682508"} Oct 03 15:02:57 crc kubenswrapper[4959]: I1003 15:02:57.612291 4959 generic.go:334] "Generic (PLEG): container finished" podID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerID="d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92" exitCode=0 Oct 03 15:02:57 crc kubenswrapper[4959]: I1003 15:02:57.612371 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerDied","Data":"d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92"} Oct 03 15:02:58 crc kubenswrapper[4959]: I1003 15:02:58.623828 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerStarted","Data":"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2"} Oct 03 15:02:58 crc kubenswrapper[4959]: I1003 15:02:58.654252 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fwbt7" podStartSLOduration=3.189812607 podStartE2EDuration="4.65423144s" podCreationTimestamp="2025-10-03 15:02:54 +0000 UTC" firstStartedPulling="2025-10-03 15:02:56.603622054 +0000 UTC m=+5545.806965481" lastFinishedPulling="2025-10-03 15:02:58.068040897 +0000 UTC m=+5547.271384314" observedRunningTime="2025-10-03 15:02:58.648285795 +0000 UTC m=+5547.851629202" watchObservedRunningTime="2025-10-03 15:02:58.65423144 +0000 UTC m=+5547.857574857" Oct 03 15:03:05 crc kubenswrapper[4959]: I1003 15:03:05.364767 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:05 crc kubenswrapper[4959]: I1003 15:03:05.365292 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:05 crc kubenswrapper[4959]: I1003 15:03:05.413386 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:05 crc kubenswrapper[4959]: I1003 15:03:05.732851 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:05 crc kubenswrapper[4959]: I1003 15:03:05.785907 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:03:07 crc kubenswrapper[4959]: I1003 15:03:07.721586 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fwbt7" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="registry-server" containerID="cri-o://a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2" gracePeriod=2 Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.188471 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.315259 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities\") pod \"a205c470-f37a-4a98-92a6-0dca74424d9d\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.315376 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28kl5\" (UniqueName: \"kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5\") pod \"a205c470-f37a-4a98-92a6-0dca74424d9d\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.315446 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content\") pod \"a205c470-f37a-4a98-92a6-0dca74424d9d\" (UID: \"a205c470-f37a-4a98-92a6-0dca74424d9d\") " Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.316664 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities" (OuterVolumeSpecName: "utilities") pod "a205c470-f37a-4a98-92a6-0dca74424d9d" (UID: "a205c470-f37a-4a98-92a6-0dca74424d9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.322590 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5" (OuterVolumeSpecName: "kube-api-access-28kl5") pod "a205c470-f37a-4a98-92a6-0dca74424d9d" (UID: "a205c470-f37a-4a98-92a6-0dca74424d9d"). InnerVolumeSpecName "kube-api-access-28kl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.328283 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a205c470-f37a-4a98-92a6-0dca74424d9d" (UID: "a205c470-f37a-4a98-92a6-0dca74424d9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.416869 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.416904 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a205c470-f37a-4a98-92a6-0dca74424d9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.416914 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28kl5\" (UniqueName: \"kubernetes.io/projected/a205c470-f37a-4a98-92a6-0dca74424d9d-kube-api-access-28kl5\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.739021 4959 generic.go:334] "Generic (PLEG): container finished" podID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerID="a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2" exitCode=0 Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.739075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerDied","Data":"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2"} Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.739109 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwbt7" event={"ID":"a205c470-f37a-4a98-92a6-0dca74424d9d","Type":"ContainerDied","Data":"c46034cfda9b95f8c89faa595fdd5be90ae3cd7f85156e830c1012b7c7682508"} Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.739131 4959 scope.go:117] "RemoveContainer" containerID="a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.739346 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwbt7" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.778146 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.785516 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwbt7"] Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.791351 4959 scope.go:117] "RemoveContainer" containerID="d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.813956 4959 scope.go:117] "RemoveContainer" containerID="3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.850768 4959 scope.go:117] "RemoveContainer" containerID="a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2" Oct 03 15:03:08 crc kubenswrapper[4959]: E1003 15:03:08.851153 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2\": container with ID starting with a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2 not found: ID does not exist" containerID="a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.851183 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2"} err="failed to get container status \"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2\": rpc error: code = NotFound desc = could not find container \"a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2\": container with ID starting with a2e816e077c1c101632de687dccbf43eba063a44fabd71be198ec4e1805eb6e2 not found: ID does not exist" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.851216 4959 scope.go:117] "RemoveContainer" containerID="d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92" Oct 03 15:03:08 crc kubenswrapper[4959]: E1003 15:03:08.851576 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92\": container with ID starting with d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92 not found: ID does not exist" containerID="d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.851594 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92"} err="failed to get container status \"d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92\": rpc error: code = NotFound desc = could not find container \"d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92\": container with ID starting with d16b23110aaedeeb2ac8e64aa0cec5e017b59008460b007e671095142c7b7e92 not found: ID does not exist" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.851714 4959 scope.go:117] "RemoveContainer" containerID="3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b" Oct 03 15:03:08 crc kubenswrapper[4959]: E1003 15:03:08.852249 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b\": container with ID starting with 3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b not found: ID does not exist" containerID="3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b" Oct 03 15:03:08 crc kubenswrapper[4959]: I1003 15:03:08.852272 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b"} err="failed to get container status \"3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b\": rpc error: code = NotFound desc = could not find container \"3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b\": container with ID starting with 3c14e08657ecfef30039b70b6d6fdca5ecc0a0dcd5f5752b30b4e61ef1b0b32b not found: ID does not exist" Oct 03 15:03:09 crc kubenswrapper[4959]: I1003 15:03:09.703151 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" path="/var/lib/kubelet/pods/a205c470-f37a-4a98-92a6-0dca74424d9d/volumes" Oct 03 15:03:13 crc kubenswrapper[4959]: E1003 15:03:13.669052 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:56526->38.129.56.245:39745: write tcp 38.129.56.245:56526->38.129.56.245:39745: write: broken pipe Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.539393 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8l4kz"] Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.539981 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="extract-content" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.539993 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="extract-content" Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.540001 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540007 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.540035 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540042 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.540055 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="extract-content" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540060 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="extract-content" Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.540070 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="extract-utilities" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540076 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="extract-utilities" Oct 03 15:03:15 crc kubenswrapper[4959]: E1003 15:03:15.540089 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="extract-utilities" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540095 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="extract-utilities" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540311 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a205c470-f37a-4a98-92a6-0dca74424d9d" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.540339 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="597deac7-221b-4cd6-88fe-d7d628790248" containerName="registry-server" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.541031 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.552644 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8l4kz"] Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.560036 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7pq8\" (UniqueName: \"kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8\") pod \"barbican-db-create-8l4kz\" (UID: \"89d83140-12ff-4a67-83df-5d645515c275\") " pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.661391 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7pq8\" (UniqueName: \"kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8\") pod \"barbican-db-create-8l4kz\" (UID: \"89d83140-12ff-4a67-83df-5d645515c275\") " pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.686080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7pq8\" (UniqueName: \"kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8\") pod \"barbican-db-create-8l4kz\" (UID: \"89d83140-12ff-4a67-83df-5d645515c275\") " pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:15 crc kubenswrapper[4959]: I1003 15:03:15.869262 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:16 crc kubenswrapper[4959]: I1003 15:03:16.339485 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8l4kz"] Oct 03 15:03:16 crc kubenswrapper[4959]: I1003 15:03:16.822237 4959 generic.go:334] "Generic (PLEG): container finished" podID="89d83140-12ff-4a67-83df-5d645515c275" containerID="ee9ddde554b4e72117583d47ba1873bcfe4b4eaedebf8cd77a1c50d7d26c1e4a" exitCode=0 Oct 03 15:03:16 crc kubenswrapper[4959]: I1003 15:03:16.822347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8l4kz" event={"ID":"89d83140-12ff-4a67-83df-5d645515c275","Type":"ContainerDied","Data":"ee9ddde554b4e72117583d47ba1873bcfe4b4eaedebf8cd77a1c50d7d26c1e4a"} Oct 03 15:03:16 crc kubenswrapper[4959]: I1003 15:03:16.822583 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8l4kz" event={"ID":"89d83140-12ff-4a67-83df-5d645515c275","Type":"ContainerStarted","Data":"cdb0f67bd3385246db00b108d0c87d4243b469df1eeb9edef897909752ed372d"} Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.200747 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.312646 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7pq8\" (UniqueName: \"kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8\") pod \"89d83140-12ff-4a67-83df-5d645515c275\" (UID: \"89d83140-12ff-4a67-83df-5d645515c275\") " Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.318724 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8" (OuterVolumeSpecName: "kube-api-access-j7pq8") pod "89d83140-12ff-4a67-83df-5d645515c275" (UID: "89d83140-12ff-4a67-83df-5d645515c275"). InnerVolumeSpecName "kube-api-access-j7pq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.414808 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7pq8\" (UniqueName: \"kubernetes.io/projected/89d83140-12ff-4a67-83df-5d645515c275-kube-api-access-j7pq8\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.845215 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8l4kz" event={"ID":"89d83140-12ff-4a67-83df-5d645515c275","Type":"ContainerDied","Data":"cdb0f67bd3385246db00b108d0c87d4243b469df1eeb9edef897909752ed372d"} Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.845270 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdb0f67bd3385246db00b108d0c87d4243b469df1eeb9edef897909752ed372d" Oct 03 15:03:18 crc kubenswrapper[4959]: I1003 15:03:18.845349 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8l4kz" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.661525 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8a79-account-create-pqb8q"] Oct 03 15:03:25 crc kubenswrapper[4959]: E1003 15:03:25.662467 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d83140-12ff-4a67-83df-5d645515c275" containerName="mariadb-database-create" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.662484 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d83140-12ff-4a67-83df-5d645515c275" containerName="mariadb-database-create" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.662701 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d83140-12ff-4a67-83df-5d645515c275" containerName="mariadb-database-create" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.663380 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.666519 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.672860 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a79-account-create-pqb8q"] Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.849594 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wlmt\" (UniqueName: \"kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt\") pod \"barbican-8a79-account-create-pqb8q\" (UID: \"76be3406-5b1b-47a6-990e-1c28f52e9ada\") " pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.951063 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wlmt\" (UniqueName: \"kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt\") pod \"barbican-8a79-account-create-pqb8q\" (UID: \"76be3406-5b1b-47a6-990e-1c28f52e9ada\") " pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.977117 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wlmt\" (UniqueName: \"kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt\") pod \"barbican-8a79-account-create-pqb8q\" (UID: \"76be3406-5b1b-47a6-990e-1c28f52e9ada\") " pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:25 crc kubenswrapper[4959]: I1003 15:03:25.986923 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:26 crc kubenswrapper[4959]: I1003 15:03:26.446399 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a79-account-create-pqb8q"] Oct 03 15:03:26 crc kubenswrapper[4959]: I1003 15:03:26.926051 4959 generic.go:334] "Generic (PLEG): container finished" podID="76be3406-5b1b-47a6-990e-1c28f52e9ada" containerID="04f3cb5cf8e17e03130963fa7fb750bee7ad94650c21e349f56637d1c63e0fd7" exitCode=0 Oct 03 15:03:26 crc kubenswrapper[4959]: I1003 15:03:26.926440 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a79-account-create-pqb8q" event={"ID":"76be3406-5b1b-47a6-990e-1c28f52e9ada","Type":"ContainerDied","Data":"04f3cb5cf8e17e03130963fa7fb750bee7ad94650c21e349f56637d1c63e0fd7"} Oct 03 15:03:26 crc kubenswrapper[4959]: I1003 15:03:26.926474 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a79-account-create-pqb8q" event={"ID":"76be3406-5b1b-47a6-990e-1c28f52e9ada","Type":"ContainerStarted","Data":"f45a074f3b171750fdcb56fddf235b1310cffe0838e1326a055155d7ac543042"} Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.300164 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.397645 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wlmt\" (UniqueName: \"kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt\") pod \"76be3406-5b1b-47a6-990e-1c28f52e9ada\" (UID: \"76be3406-5b1b-47a6-990e-1c28f52e9ada\") " Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.410420 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt" (OuterVolumeSpecName: "kube-api-access-4wlmt") pod "76be3406-5b1b-47a6-990e-1c28f52e9ada" (UID: "76be3406-5b1b-47a6-990e-1c28f52e9ada"). InnerVolumeSpecName "kube-api-access-4wlmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.499828 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wlmt\" (UniqueName: \"kubernetes.io/projected/76be3406-5b1b-47a6-990e-1c28f52e9ada-kube-api-access-4wlmt\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.947118 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a79-account-create-pqb8q" event={"ID":"76be3406-5b1b-47a6-990e-1c28f52e9ada","Type":"ContainerDied","Data":"f45a074f3b171750fdcb56fddf235b1310cffe0838e1326a055155d7ac543042"} Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.947167 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f45a074f3b171750fdcb56fddf235b1310cffe0838e1326a055155d7ac543042" Oct 03 15:03:28 crc kubenswrapper[4959]: I1003 15:03:28.947176 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a79-account-create-pqb8q" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.950699 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-prmj8"] Oct 03 15:03:30 crc kubenswrapper[4959]: E1003 15:03:30.951464 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76be3406-5b1b-47a6-990e-1c28f52e9ada" containerName="mariadb-account-create" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.951481 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="76be3406-5b1b-47a6-990e-1c28f52e9ada" containerName="mariadb-account-create" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.951724 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="76be3406-5b1b-47a6-990e-1c28f52e9ada" containerName="mariadb-account-create" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.952438 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.955559 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-l9gzg" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.956085 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 15:03:30 crc kubenswrapper[4959]: I1003 15:03:30.962499 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-prmj8"] Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.046360 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.046438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.046459 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx68t\" (UniqueName: \"kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.148510 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.148607 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.148631 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx68t\" (UniqueName: \"kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.156364 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.160725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.171443 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx68t\" (UniqueName: \"kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t\") pod \"barbican-db-sync-prmj8\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.284464 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.760096 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-prmj8"] Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.976726 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-prmj8" event={"ID":"a7acf8ef-a0b7-4f3a-a180-0c3f56580328","Type":"ContainerStarted","Data":"367703c6fd0a2b36361ffd3982b08910634276352fc8f1f7b635b00e716286ad"} Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.976776 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-prmj8" event={"ID":"a7acf8ef-a0b7-4f3a-a180-0c3f56580328","Type":"ContainerStarted","Data":"6bbd3af3a291934fa2644448a559d66bce149194c7b18388eeef7107a113c4be"} Oct 03 15:03:31 crc kubenswrapper[4959]: I1003 15:03:31.993904 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-prmj8" podStartSLOduration=1.993877233 podStartE2EDuration="1.993877233s" podCreationTimestamp="2025-10-03 15:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:03:31.992567861 +0000 UTC m=+5581.195911278" watchObservedRunningTime="2025-10-03 15:03:31.993877233 +0000 UTC m=+5581.197220670" Oct 03 15:03:33 crc kubenswrapper[4959]: I1003 15:03:33.995028 4959 generic.go:334] "Generic (PLEG): container finished" podID="a7acf8ef-a0b7-4f3a-a180-0c3f56580328" containerID="367703c6fd0a2b36361ffd3982b08910634276352fc8f1f7b635b00e716286ad" exitCode=0 Oct 03 15:03:33 crc kubenswrapper[4959]: I1003 15:03:33.995088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-prmj8" event={"ID":"a7acf8ef-a0b7-4f3a-a180-0c3f56580328","Type":"ContainerDied","Data":"367703c6fd0a2b36361ffd3982b08910634276352fc8f1f7b635b00e716286ad"} Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.352782 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.425159 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx68t\" (UniqueName: \"kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t\") pod \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.425524 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle\") pod \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.425658 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data\") pod \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\" (UID: \"a7acf8ef-a0b7-4f3a-a180-0c3f56580328\") " Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.431928 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a7acf8ef-a0b7-4f3a-a180-0c3f56580328" (UID: "a7acf8ef-a0b7-4f3a-a180-0c3f56580328"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.433183 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t" (OuterVolumeSpecName: "kube-api-access-vx68t") pod "a7acf8ef-a0b7-4f3a-a180-0c3f56580328" (UID: "a7acf8ef-a0b7-4f3a-a180-0c3f56580328"). InnerVolumeSpecName "kube-api-access-vx68t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.450466 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7acf8ef-a0b7-4f3a-a180-0c3f56580328" (UID: "a7acf8ef-a0b7-4f3a-a180-0c3f56580328"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.528315 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.528440 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx68t\" (UniqueName: \"kubernetes.io/projected/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-kube-api-access-vx68t\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:35 crc kubenswrapper[4959]: I1003 15:03:35.528473 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7acf8ef-a0b7-4f3a-a180-0c3f56580328-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.016830 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-prmj8" event={"ID":"a7acf8ef-a0b7-4f3a-a180-0c3f56580328","Type":"ContainerDied","Data":"6bbd3af3a291934fa2644448a559d66bce149194c7b18388eeef7107a113c4be"} Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.017179 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bbd3af3a291934fa2644448a559d66bce149194c7b18388eeef7107a113c4be" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.017347 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-prmj8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.246112 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6d6558856b-h9zfg"] Oct 03 15:03:36 crc kubenswrapper[4959]: E1003 15:03:36.246550 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7acf8ef-a0b7-4f3a-a180-0c3f56580328" containerName="barbican-db-sync" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.246569 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7acf8ef-a0b7-4f3a-a180-0c3f56580328" containerName="barbican-db-sync" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.246738 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7acf8ef-a0b7-4f3a-a180-0c3f56580328" containerName="barbican-db-sync" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.247721 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.252518 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-l9gzg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.252774 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.254523 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.258300 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-d48b6b55f-s88sc"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.259658 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.261073 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.273529 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d6558856b-h9zfg"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.284618 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d48b6b55f-s88sc"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.322867 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.325275 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.337841 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340807 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv97m\" (UniqueName: \"kubernetes.io/projected/217643a2-9b13-4460-b9bd-7e01a2235174-kube-api-access-zv97m\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-combined-ca-bundle\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340877 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340892 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e10ff4ef-9f81-416c-a313-91c8313d2104-logs\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340925 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340949 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xppmg\" (UniqueName: \"kubernetes.io/projected/e10ff4ef-9f81-416c-a313-91c8313d2104-kube-api-access-xppmg\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.340973 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217643a2-9b13-4460-b9bd-7e01a2235174-logs\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.341002 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data-custom\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.341050 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.341076 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data-custom\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.438326 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f4b8ff658-pvrk8"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.439612 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442073 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bclkf\" (UniqueName: \"kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442103 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217643a2-9b13-4460-b9bd-7e01a2235174-logs\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442137 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data-custom\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442243 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data-custom\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442334 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv97m\" (UniqueName: \"kubernetes.io/projected/217643a2-9b13-4460-b9bd-7e01a2235174-kube-api-access-zv97m\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442359 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-combined-ca-bundle\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e10ff4ef-9f81-416c-a313-91c8313d2104-logs\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442414 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442445 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442470 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.442495 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xppmg\" (UniqueName: \"kubernetes.io/projected/e10ff4ef-9f81-416c-a313-91c8313d2104-kube-api-access-xppmg\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.443185 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217643a2-9b13-4460-b9bd-7e01a2235174-logs\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.444482 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.445725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e10ff4ef-9f81-416c-a313-91c8313d2104-logs\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.450028 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data-custom\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.450466 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-combined-ca-bundle\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.452755 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f4b8ff658-pvrk8"] Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.455354 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.456573 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-config-data\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.457074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/217643a2-9b13-4460-b9bd-7e01a2235174-config-data-custom\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.459845 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e10ff4ef-9f81-416c-a313-91c8313d2104-combined-ca-bundle\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.465111 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xppmg\" (UniqueName: \"kubernetes.io/projected/e10ff4ef-9f81-416c-a313-91c8313d2104-kube-api-access-xppmg\") pod \"barbican-worker-d48b6b55f-s88sc\" (UID: \"e10ff4ef-9f81-416c-a313-91c8313d2104\") " pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.466647 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv97m\" (UniqueName: \"kubernetes.io/projected/217643a2-9b13-4460-b9bd-7e01a2235174-kube-api-access-zv97m\") pod \"barbican-keystone-listener-6d6558856b-h9zfg\" (UID: \"217643a2-9b13-4460-b9bd-7e01a2235174\") " pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.544484 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.544816 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data-custom\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.544934 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545052 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545162 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bclkf\" (UniqueName: \"kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545288 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545370 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545467 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb84w\" (UniqueName: \"kubernetes.io/projected/e85cb5d9-0483-4ebf-979f-c614249eaa4d-kube-api-access-hb84w\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545555 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-combined-ca-bundle\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545759 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e85cb5d9-0483-4ebf-979f-c614249eaa4d-logs\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545849 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.545849 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.546480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.546577 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.563599 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bclkf\" (UniqueName: \"kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf\") pod \"dnsmasq-dns-596df78cd9-cwj4s\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.570451 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.581689 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d48b6b55f-s88sc" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.646752 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb84w\" (UniqueName: \"kubernetes.io/projected/e85cb5d9-0483-4ebf-979f-c614249eaa4d-kube-api-access-hb84w\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.646796 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-combined-ca-bundle\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.646841 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e85cb5d9-0483-4ebf-979f-c614249eaa4d-logs\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.646889 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data-custom\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.646937 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.647643 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.648615 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e85cb5d9-0483-4ebf-979f-c614249eaa4d-logs\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.651708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-combined-ca-bundle\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.654700 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data-custom\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.657166 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e85cb5d9-0483-4ebf-979f-c614249eaa4d-config-data\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.664016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb84w\" (UniqueName: \"kubernetes.io/projected/e85cb5d9-0483-4ebf-979f-c614249eaa4d-kube-api-access-hb84w\") pod \"barbican-api-7f4b8ff658-pvrk8\" (UID: \"e85cb5d9-0483-4ebf-979f-c614249eaa4d\") " pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:36 crc kubenswrapper[4959]: I1003 15:03:36.823417 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:37 crc kubenswrapper[4959]: I1003 15:03:37.108252 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d48b6b55f-s88sc"] Oct 03 15:03:37 crc kubenswrapper[4959]: W1003 15:03:37.123173 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode10ff4ef_9f81_416c_a313_91c8313d2104.slice/crio-ad09388d0890d9dab632c88bd29493a11578ee9e8c2364aa789bf1ac759269ae WatchSource:0}: Error finding container ad09388d0890d9dab632c88bd29493a11578ee9e8c2364aa789bf1ac759269ae: Status 404 returned error can't find the container with id ad09388d0890d9dab632c88bd29493a11578ee9e8c2364aa789bf1ac759269ae Oct 03 15:03:37 crc kubenswrapper[4959]: I1003 15:03:37.191319 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d6558856b-h9zfg"] Oct 03 15:03:37 crc kubenswrapper[4959]: W1003 15:03:37.200505 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod217643a2_9b13_4460_b9bd_7e01a2235174.slice/crio-c42205cc6dbc5a2f81ebcf5a9e659ae1a8e435ffe714b81a3efb670e04d9e8dd WatchSource:0}: Error finding container c42205cc6dbc5a2f81ebcf5a9e659ae1a8e435ffe714b81a3efb670e04d9e8dd: Status 404 returned error can't find the container with id c42205cc6dbc5a2f81ebcf5a9e659ae1a8e435ffe714b81a3efb670e04d9e8dd Oct 03 15:03:37 crc kubenswrapper[4959]: I1003 15:03:37.267388 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:03:37 crc kubenswrapper[4959]: W1003 15:03:37.272910 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod012bdeaa_9454_4893_996c_0d231f323c43.slice/crio-c257ad4b7f6b89786b152b12ec0d9e05b58fd4ca12f00afa149727fdc5940075 WatchSource:0}: Error finding container c257ad4b7f6b89786b152b12ec0d9e05b58fd4ca12f00afa149727fdc5940075: Status 404 returned error can't find the container with id c257ad4b7f6b89786b152b12ec0d9e05b58fd4ca12f00afa149727fdc5940075 Oct 03 15:03:37 crc kubenswrapper[4959]: I1003 15:03:37.343502 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f4b8ff658-pvrk8"] Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.034898 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f4b8ff658-pvrk8" event={"ID":"e85cb5d9-0483-4ebf-979f-c614249eaa4d","Type":"ContainerStarted","Data":"2edc7515a489fc12dfa68fc3d8251aa261304c44bb03ce41c6496281034d15ca"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.035267 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f4b8ff658-pvrk8" event={"ID":"e85cb5d9-0483-4ebf-979f-c614249eaa4d","Type":"ContainerStarted","Data":"8cbdaff4896e9cda24f0bf676954564eccdc037be9174639b29800053b6ac7d5"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.037538 4959 generic.go:334] "Generic (PLEG): container finished" podID="012bdeaa-9454-4893-996c-0d231f323c43" containerID="56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d" exitCode=0 Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.037616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" event={"ID":"012bdeaa-9454-4893-996c-0d231f323c43","Type":"ContainerDied","Data":"56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.037638 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" event={"ID":"012bdeaa-9454-4893-996c-0d231f323c43","Type":"ContainerStarted","Data":"c257ad4b7f6b89786b152b12ec0d9e05b58fd4ca12f00afa149727fdc5940075"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.042218 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d48b6b55f-s88sc" event={"ID":"e10ff4ef-9f81-416c-a313-91c8313d2104","Type":"ContainerStarted","Data":"78ed4d8036af85ef83b73b8cea03595e02c3c0ba90e2ce3333ac033770a90d8a"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.042252 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d48b6b55f-s88sc" event={"ID":"e10ff4ef-9f81-416c-a313-91c8313d2104","Type":"ContainerStarted","Data":"067a3a9ff58d0d99f1bde886525caef66d054788c056e449c2c41f3ace139d2f"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.042266 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d48b6b55f-s88sc" event={"ID":"e10ff4ef-9f81-416c-a313-91c8313d2104","Type":"ContainerStarted","Data":"ad09388d0890d9dab632c88bd29493a11578ee9e8c2364aa789bf1ac759269ae"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.050781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" event={"ID":"217643a2-9b13-4460-b9bd-7e01a2235174","Type":"ContainerStarted","Data":"28652d188ca4c1289f19dd698c995013ff89dc09b1e874d64cc8f1d26fbe3ed7"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.050850 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" event={"ID":"217643a2-9b13-4460-b9bd-7e01a2235174","Type":"ContainerStarted","Data":"491c01f328b962669d8ed090aa02c408a3738fd2473cb2abcd601eb51eb6a573"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.050863 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" event={"ID":"217643a2-9b13-4460-b9bd-7e01a2235174","Type":"ContainerStarted","Data":"c42205cc6dbc5a2f81ebcf5a9e659ae1a8e435ffe714b81a3efb670e04d9e8dd"} Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.081739 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6d6558856b-h9zfg" podStartSLOduration=2.081706386 podStartE2EDuration="2.081706386s" podCreationTimestamp="2025-10-03 15:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:03:38.075498454 +0000 UTC m=+5587.278841871" watchObservedRunningTime="2025-10-03 15:03:38.081706386 +0000 UTC m=+5587.285049813" Oct 03 15:03:38 crc kubenswrapper[4959]: I1003 15:03:38.103114 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-d48b6b55f-s88sc" podStartSLOduration=2.102935536 podStartE2EDuration="2.102935536s" podCreationTimestamp="2025-10-03 15:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:03:38.096563119 +0000 UTC m=+5587.299906556" watchObservedRunningTime="2025-10-03 15:03:38.102935536 +0000 UTC m=+5587.306278953" Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.058812 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f4b8ff658-pvrk8" event={"ID":"e85cb5d9-0483-4ebf-979f-c614249eaa4d","Type":"ContainerStarted","Data":"ace3ab1852a834b2256cf6607b439ee722c728ad0b7b4cb378c48522a7aba7aa"} Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.059131 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.059152 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.061897 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" event={"ID":"012bdeaa-9454-4893-996c-0d231f323c43","Type":"ContainerStarted","Data":"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5"} Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.061948 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.082146 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f4b8ff658-pvrk8" podStartSLOduration=3.082125235 podStartE2EDuration="3.082125235s" podCreationTimestamp="2025-10-03 15:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:03:39.078138367 +0000 UTC m=+5588.281481854" watchObservedRunningTime="2025-10-03 15:03:39.082125235 +0000 UTC m=+5588.285468662" Oct 03 15:03:39 crc kubenswrapper[4959]: I1003 15:03:39.108851 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" podStartSLOduration=3.108823138 podStartE2EDuration="3.108823138s" podCreationTimestamp="2025-10-03 15:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:03:39.099529861 +0000 UTC m=+5588.302873278" watchObservedRunningTime="2025-10-03 15:03:39.108823138 +0000 UTC m=+5588.312166585" Oct 03 15:03:46 crc kubenswrapper[4959]: I1003 15:03:46.650495 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:03:46 crc kubenswrapper[4959]: I1003 15:03:46.728164 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:03:46 crc kubenswrapper[4959]: I1003 15:03:46.728619 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58865cd75-fsg97" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="dnsmasq-dns" containerID="cri-o://7c0e5369c94b466222957921c028bb46715f6cc1b79e946d41f24becd89c5b6f" gracePeriod=10 Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.135052 4959 generic.go:334] "Generic (PLEG): container finished" podID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerID="7c0e5369c94b466222957921c028bb46715f6cc1b79e946d41f24becd89c5b6f" exitCode=0 Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.135122 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58865cd75-fsg97" event={"ID":"e59377d0-0346-4d14-8b23-2e5d88a8e6ff","Type":"ContainerDied","Data":"7c0e5369c94b466222957921c028bb46715f6cc1b79e946d41f24becd89c5b6f"} Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.225492 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.366238 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb\") pod \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.366652 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb\") pod \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.366671 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config\") pod \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.366712 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc\") pod \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.366791 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lhqx\" (UniqueName: \"kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx\") pod \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\" (UID: \"e59377d0-0346-4d14-8b23-2e5d88a8e6ff\") " Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.402938 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx" (OuterVolumeSpecName: "kube-api-access-9lhqx") pod "e59377d0-0346-4d14-8b23-2e5d88a8e6ff" (UID: "e59377d0-0346-4d14-8b23-2e5d88a8e6ff"). InnerVolumeSpecName "kube-api-access-9lhqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.469051 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lhqx\" (UniqueName: \"kubernetes.io/projected/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-kube-api-access-9lhqx\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.488168 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e59377d0-0346-4d14-8b23-2e5d88a8e6ff" (UID: "e59377d0-0346-4d14-8b23-2e5d88a8e6ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.495852 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e59377d0-0346-4d14-8b23-2e5d88a8e6ff" (UID: "e59377d0-0346-4d14-8b23-2e5d88a8e6ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.536244 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config" (OuterVolumeSpecName: "config") pod "e59377d0-0346-4d14-8b23-2e5d88a8e6ff" (UID: "e59377d0-0346-4d14-8b23-2e5d88a8e6ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.542591 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e59377d0-0346-4d14-8b23-2e5d88a8e6ff" (UID: "e59377d0-0346-4d14-8b23-2e5d88a8e6ff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.571543 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.571586 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.571600 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:47 crc kubenswrapper[4959]: I1003 15:03:47.571610 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e59377d0-0346-4d14-8b23-2e5d88a8e6ff-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.147758 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58865cd75-fsg97" event={"ID":"e59377d0-0346-4d14-8b23-2e5d88a8e6ff","Type":"ContainerDied","Data":"c9f89c93092104209a856dda973237ecbf4febf1e67a73c0500f0491a26b739e"} Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.148963 4959 scope.go:117] "RemoveContainer" containerID="7c0e5369c94b466222957921c028bb46715f6cc1b79e946d41f24becd89c5b6f" Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.149249 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58865cd75-fsg97" Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.184853 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.197407 4959 scope.go:117] "RemoveContainer" containerID="cd70d4f50dee87d8663aa8e7587d8d3480641b2fd80243f37604b057716d908a" Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.202560 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58865cd75-fsg97"] Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.494446 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:48 crc kubenswrapper[4959]: I1003 15:03:48.514711 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f4b8ff658-pvrk8" Oct 03 15:03:49 crc kubenswrapper[4959]: I1003 15:03:49.702644 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" path="/var/lib/kubelet/pods/e59377d0-0346-4d14-8b23-2e5d88a8e6ff/volumes" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.675275 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-npfx5"] Oct 03 15:04:02 crc kubenswrapper[4959]: E1003 15:04:02.676394 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="dnsmasq-dns" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.676414 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="dnsmasq-dns" Oct 03 15:04:02 crc kubenswrapper[4959]: E1003 15:04:02.676441 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="init" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.676449 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="init" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.676649 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59377d0-0346-4d14-8b23-2e5d88a8e6ff" containerName="dnsmasq-dns" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.677366 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.685791 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-npfx5"] Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.789758 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr5z2\" (UniqueName: \"kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2\") pod \"neutron-db-create-npfx5\" (UID: \"e14283b9-fe65-49a2-ac39-43429bb24a4e\") " pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.892311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr5z2\" (UniqueName: \"kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2\") pod \"neutron-db-create-npfx5\" (UID: \"e14283b9-fe65-49a2-ac39-43429bb24a4e\") " pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:02 crc kubenswrapper[4959]: I1003 15:04:02.912536 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr5z2\" (UniqueName: \"kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2\") pod \"neutron-db-create-npfx5\" (UID: \"e14283b9-fe65-49a2-ac39-43429bb24a4e\") " pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:03 crc kubenswrapper[4959]: I1003 15:04:03.000666 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:03 crc kubenswrapper[4959]: I1003 15:04:03.548941 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-npfx5"] Oct 03 15:04:04 crc kubenswrapper[4959]: I1003 15:04:04.297282 4959 generic.go:334] "Generic (PLEG): container finished" podID="e14283b9-fe65-49a2-ac39-43429bb24a4e" containerID="097bf3638a3d687477aea94d3077f7ed9a8248820dbb5b43bd3d78a34fe62d3b" exitCode=0 Oct 03 15:04:04 crc kubenswrapper[4959]: I1003 15:04:04.297387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-npfx5" event={"ID":"e14283b9-fe65-49a2-ac39-43429bb24a4e","Type":"ContainerDied","Data":"097bf3638a3d687477aea94d3077f7ed9a8248820dbb5b43bd3d78a34fe62d3b"} Oct 03 15:04:04 crc kubenswrapper[4959]: I1003 15:04:04.297655 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-npfx5" event={"ID":"e14283b9-fe65-49a2-ac39-43429bb24a4e","Type":"ContainerStarted","Data":"627120b37823896b19f2678433cd0bf94f06e9dd6730093d596b11c5494b8430"} Oct 03 15:04:05 crc kubenswrapper[4959]: I1003 15:04:05.684116 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:05 crc kubenswrapper[4959]: I1003 15:04:05.747263 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr5z2\" (UniqueName: \"kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2\") pod \"e14283b9-fe65-49a2-ac39-43429bb24a4e\" (UID: \"e14283b9-fe65-49a2-ac39-43429bb24a4e\") " Oct 03 15:04:05 crc kubenswrapper[4959]: I1003 15:04:05.756622 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2" (OuterVolumeSpecName: "kube-api-access-fr5z2") pod "e14283b9-fe65-49a2-ac39-43429bb24a4e" (UID: "e14283b9-fe65-49a2-ac39-43429bb24a4e"). InnerVolumeSpecName "kube-api-access-fr5z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:04:05 crc kubenswrapper[4959]: I1003 15:04:05.849700 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr5z2\" (UniqueName: \"kubernetes.io/projected/e14283b9-fe65-49a2-ac39-43429bb24a4e-kube-api-access-fr5z2\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:06 crc kubenswrapper[4959]: I1003 15:04:06.044677 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:04:06 crc kubenswrapper[4959]: I1003 15:04:06.044799 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:04:06 crc kubenswrapper[4959]: I1003 15:04:06.333564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-npfx5" event={"ID":"e14283b9-fe65-49a2-ac39-43429bb24a4e","Type":"ContainerDied","Data":"627120b37823896b19f2678433cd0bf94f06e9dd6730093d596b11c5494b8430"} Oct 03 15:04:06 crc kubenswrapper[4959]: I1003 15:04:06.333633 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="627120b37823896b19f2678433cd0bf94f06e9dd6730093d596b11c5494b8430" Oct 03 15:04:06 crc kubenswrapper[4959]: I1003 15:04:06.333639 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-npfx5" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.785289 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6131-account-create-gmvbg"] Oct 03 15:04:12 crc kubenswrapper[4959]: E1003 15:04:12.786134 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14283b9-fe65-49a2-ac39-43429bb24a4e" containerName="mariadb-database-create" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.786152 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14283b9-fe65-49a2-ac39-43429bb24a4e" containerName="mariadb-database-create" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.786482 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14283b9-fe65-49a2-ac39-43429bb24a4e" containerName="mariadb-database-create" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.789436 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.791810 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.794907 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6131-account-create-gmvbg"] Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.886550 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds54g\" (UniqueName: \"kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g\") pod \"neutron-6131-account-create-gmvbg\" (UID: \"8a15d14e-bbad-4071-8531-ee330e283a2a\") " pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:12 crc kubenswrapper[4959]: I1003 15:04:12.988880 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds54g\" (UniqueName: \"kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g\") pod \"neutron-6131-account-create-gmvbg\" (UID: \"8a15d14e-bbad-4071-8531-ee330e283a2a\") " pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:13 crc kubenswrapper[4959]: I1003 15:04:13.014153 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds54g\" (UniqueName: \"kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g\") pod \"neutron-6131-account-create-gmvbg\" (UID: \"8a15d14e-bbad-4071-8531-ee330e283a2a\") " pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:13 crc kubenswrapper[4959]: I1003 15:04:13.111687 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:13 crc kubenswrapper[4959]: I1003 15:04:13.592370 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6131-account-create-gmvbg"] Oct 03 15:04:14 crc kubenswrapper[4959]: I1003 15:04:14.418156 4959 generic.go:334] "Generic (PLEG): container finished" podID="8a15d14e-bbad-4071-8531-ee330e283a2a" containerID="e43673f3a56fb8a117fe1293e702428f6566e2d8f81258b79981b349fcd1d50f" exitCode=0 Oct 03 15:04:14 crc kubenswrapper[4959]: I1003 15:04:14.418351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6131-account-create-gmvbg" event={"ID":"8a15d14e-bbad-4071-8531-ee330e283a2a","Type":"ContainerDied","Data":"e43673f3a56fb8a117fe1293e702428f6566e2d8f81258b79981b349fcd1d50f"} Oct 03 15:04:14 crc kubenswrapper[4959]: I1003 15:04:14.418611 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6131-account-create-gmvbg" event={"ID":"8a15d14e-bbad-4071-8531-ee330e283a2a","Type":"ContainerStarted","Data":"6f53bc9293ce4de0994128368937452a27784f92578c0c7d4dd70244e6e2ca0c"} Oct 03 15:04:15 crc kubenswrapper[4959]: I1003 15:04:15.750932 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:15 crc kubenswrapper[4959]: I1003 15:04:15.844957 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds54g\" (UniqueName: \"kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g\") pod \"8a15d14e-bbad-4071-8531-ee330e283a2a\" (UID: \"8a15d14e-bbad-4071-8531-ee330e283a2a\") " Oct 03 15:04:15 crc kubenswrapper[4959]: I1003 15:04:15.849830 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g" (OuterVolumeSpecName: "kube-api-access-ds54g") pod "8a15d14e-bbad-4071-8531-ee330e283a2a" (UID: "8a15d14e-bbad-4071-8531-ee330e283a2a"). InnerVolumeSpecName "kube-api-access-ds54g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:04:15 crc kubenswrapper[4959]: I1003 15:04:15.947109 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds54g\" (UniqueName: \"kubernetes.io/projected/8a15d14e-bbad-4071-8531-ee330e283a2a-kube-api-access-ds54g\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:16 crc kubenswrapper[4959]: I1003 15:04:16.439268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6131-account-create-gmvbg" event={"ID":"8a15d14e-bbad-4071-8531-ee330e283a2a","Type":"ContainerDied","Data":"6f53bc9293ce4de0994128368937452a27784f92578c0c7d4dd70244e6e2ca0c"} Oct 03 15:04:16 crc kubenswrapper[4959]: I1003 15:04:16.439329 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f53bc9293ce4de0994128368937452a27784f92578c0c7d4dd70244e6e2ca0c" Oct 03 15:04:16 crc kubenswrapper[4959]: I1003 15:04:16.439366 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6131-account-create-gmvbg" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.036081 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-c8twv"] Oct 03 15:04:18 crc kubenswrapper[4959]: E1003 15:04:18.036857 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a15d14e-bbad-4071-8531-ee330e283a2a" containerName="mariadb-account-create" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.036875 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a15d14e-bbad-4071-8531-ee330e283a2a" containerName="mariadb-account-create" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.037109 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a15d14e-bbad-4071-8531-ee330e283a2a" containerName="mariadb-account-create" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.037886 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.039465 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.039659 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7zshb" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.046503 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.048848 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-c8twv"] Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.184011 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.184413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.184615 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zffdh\" (UniqueName: \"kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.286465 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.286530 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.286648 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zffdh\" (UniqueName: \"kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.293755 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.294259 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.305071 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zffdh\" (UniqueName: \"kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh\") pod \"neutron-db-sync-c8twv\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.372592 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:18 crc kubenswrapper[4959]: I1003 15:04:18.842754 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-c8twv"] Oct 03 15:04:19 crc kubenswrapper[4959]: I1003 15:04:19.468407 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-c8twv" event={"ID":"949ca17b-cb34-493c-b135-300fb15c2e1a","Type":"ContainerStarted","Data":"d7a3dca16749d3380a201ff1a00c43a84b63e047384436423e1c4c1921fed001"} Oct 03 15:04:19 crc kubenswrapper[4959]: I1003 15:04:19.468811 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-c8twv" event={"ID":"949ca17b-cb34-493c-b135-300fb15c2e1a","Type":"ContainerStarted","Data":"fc43277c6bff70a746efbef9cef1a08bd3218d362ff2c93c097d2e5f84243b5d"} Oct 03 15:04:19 crc kubenswrapper[4959]: I1003 15:04:19.498066 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-c8twv" podStartSLOduration=1.498043796 podStartE2EDuration="1.498043796s" podCreationTimestamp="2025-10-03 15:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:04:19.495830251 +0000 UTC m=+5628.699173698" watchObservedRunningTime="2025-10-03 15:04:19.498043796 +0000 UTC m=+5628.701387223" Oct 03 15:04:23 crc kubenswrapper[4959]: I1003 15:04:23.510449 4959 generic.go:334] "Generic (PLEG): container finished" podID="949ca17b-cb34-493c-b135-300fb15c2e1a" containerID="d7a3dca16749d3380a201ff1a00c43a84b63e047384436423e1c4c1921fed001" exitCode=0 Oct 03 15:04:23 crc kubenswrapper[4959]: I1003 15:04:23.510864 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-c8twv" event={"ID":"949ca17b-cb34-493c-b135-300fb15c2e1a","Type":"ContainerDied","Data":"d7a3dca16749d3380a201ff1a00c43a84b63e047384436423e1c4c1921fed001"} Oct 03 15:04:24 crc kubenswrapper[4959]: I1003 15:04:24.900506 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.007569 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle\") pod \"949ca17b-cb34-493c-b135-300fb15c2e1a\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.008007 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config\") pod \"949ca17b-cb34-493c-b135-300fb15c2e1a\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.008031 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zffdh\" (UniqueName: \"kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh\") pod \"949ca17b-cb34-493c-b135-300fb15c2e1a\" (UID: \"949ca17b-cb34-493c-b135-300fb15c2e1a\") " Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.013725 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh" (OuterVolumeSpecName: "kube-api-access-zffdh") pod "949ca17b-cb34-493c-b135-300fb15c2e1a" (UID: "949ca17b-cb34-493c-b135-300fb15c2e1a"). InnerVolumeSpecName "kube-api-access-zffdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.031145 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "949ca17b-cb34-493c-b135-300fb15c2e1a" (UID: "949ca17b-cb34-493c-b135-300fb15c2e1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.045496 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config" (OuterVolumeSpecName: "config") pod "949ca17b-cb34-493c-b135-300fb15c2e1a" (UID: "949ca17b-cb34-493c-b135-300fb15c2e1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.110447 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.110492 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/949ca17b-cb34-493c-b135-300fb15c2e1a-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.110506 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zffdh\" (UniqueName: \"kubernetes.io/projected/949ca17b-cb34-493c-b135-300fb15c2e1a-kube-api-access-zffdh\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.534581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-c8twv" event={"ID":"949ca17b-cb34-493c-b135-300fb15c2e1a","Type":"ContainerDied","Data":"fc43277c6bff70a746efbef9cef1a08bd3218d362ff2c93c097d2e5f84243b5d"} Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.534617 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc43277c6bff70a746efbef9cef1a08bd3218d362ff2c93c097d2e5f84243b5d" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.534675 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-c8twv" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.657024 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:04:25 crc kubenswrapper[4959]: E1003 15:04:25.657369 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949ca17b-cb34-493c-b135-300fb15c2e1a" containerName="neutron-db-sync" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.657383 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="949ca17b-cb34-493c-b135-300fb15c2e1a" containerName="neutron-db-sync" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.657523 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="949ca17b-cb34-493c-b135-300fb15c2e1a" containerName="neutron-db-sync" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.658401 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.679594 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.728122 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.728154 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.728204 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4hkz\" (UniqueName: \"kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.728222 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.728237 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.831934 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.831973 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.832003 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4hkz\" (UniqueName: \"kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.832024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.832040 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.832970 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.833010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.833729 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.836993 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.858842 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4hkz\" (UniqueName: \"kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz\") pod \"dnsmasq-dns-5f59d59797-wbthw\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.887609 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f668759cc-jmsxf"] Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.889907 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.892037 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.892358 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-7zshb" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.892502 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 15:04:25 crc kubenswrapper[4959]: I1003 15:04:25.896607 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f668759cc-jmsxf"] Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.034365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-httpd-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.034415 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-combined-ca-bundle\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.034450 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.034507 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nl8m\" (UniqueName: \"kubernetes.io/projected/e4389945-7007-4c05-b044-2de169b96b46-kube-api-access-8nl8m\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.039913 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.136594 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nl8m\" (UniqueName: \"kubernetes.io/projected/e4389945-7007-4c05-b044-2de169b96b46-kube-api-access-8nl8m\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.136733 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-httpd-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.136780 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-combined-ca-bundle\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.136824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.142208 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-httpd-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.142325 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-config\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.142370 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4389945-7007-4c05-b044-2de169b96b46-combined-ca-bundle\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.161891 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nl8m\" (UniqueName: \"kubernetes.io/projected/e4389945-7007-4c05-b044-2de169b96b46-kube-api-access-8nl8m\") pod \"neutron-6f668759cc-jmsxf\" (UID: \"e4389945-7007-4c05-b044-2de169b96b46\") " pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.223829 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.526738 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.544418 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" event={"ID":"44e8681b-60fc-48b1-b9b1-80e36865fde0","Type":"ContainerStarted","Data":"c51b62d539fcb3526cb2da20ab2c86c3c4339e530f2f5dbc568fe981a48b2f8f"} Oct 03 15:04:26 crc kubenswrapper[4959]: I1003 15:04:26.755764 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f668759cc-jmsxf"] Oct 03 15:04:26 crc kubenswrapper[4959]: W1003 15:04:26.799923 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4389945_7007_4c05_b044_2de169b96b46.slice/crio-c75bee069c3d098fb47688d0451f2dfb8c943d8172822de5fd983dfcfb2eb19d WatchSource:0}: Error finding container c75bee069c3d098fb47688d0451f2dfb8c943d8172822de5fd983dfcfb2eb19d: Status 404 returned error can't find the container with id c75bee069c3d098fb47688d0451f2dfb8c943d8172822de5fd983dfcfb2eb19d Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.554653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f668759cc-jmsxf" event={"ID":"e4389945-7007-4c05-b044-2de169b96b46","Type":"ContainerStarted","Data":"14adba678fdb42a8ae4a6f738d9a730957bd8a53b1ec52c60f06cf55462f6c10"} Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.554968 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f668759cc-jmsxf" event={"ID":"e4389945-7007-4c05-b044-2de169b96b46","Type":"ContainerStarted","Data":"870c401fb8f38c6ad32537505ef1539e31d54e1bd717a4e91b213899d58a312d"} Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.554984 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f668759cc-jmsxf" event={"ID":"e4389945-7007-4c05-b044-2de169b96b46","Type":"ContainerStarted","Data":"c75bee069c3d098fb47688d0451f2dfb8c943d8172822de5fd983dfcfb2eb19d"} Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.555320 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.557443 4959 generic.go:334] "Generic (PLEG): container finished" podID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerID="c145715c157804ce34ff13794d9e659ea101155b8c5ff957ea75ef56d4465bba" exitCode=0 Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.557485 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" event={"ID":"44e8681b-60fc-48b1-b9b1-80e36865fde0","Type":"ContainerDied","Data":"c145715c157804ce34ff13794d9e659ea101155b8c5ff957ea75ef56d4465bba"} Oct 03 15:04:27 crc kubenswrapper[4959]: I1003 15:04:27.588873 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f668759cc-jmsxf" podStartSLOduration=2.588835209 podStartE2EDuration="2.588835209s" podCreationTimestamp="2025-10-03 15:04:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:04:27.584599726 +0000 UTC m=+5636.787943143" watchObservedRunningTime="2025-10-03 15:04:27.588835209 +0000 UTC m=+5636.792178626" Oct 03 15:04:28 crc kubenswrapper[4959]: I1003 15:04:28.568267 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" event={"ID":"44e8681b-60fc-48b1-b9b1-80e36865fde0","Type":"ContainerStarted","Data":"f4225ff6cb567f3b52fcf43640e9f093f08c003aa49a53a289bd097d6f67bd09"} Oct 03 15:04:28 crc kubenswrapper[4959]: I1003 15:04:28.568520 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:28 crc kubenswrapper[4959]: I1003 15:04:28.587933 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" podStartSLOduration=3.587916556 podStartE2EDuration="3.587916556s" podCreationTimestamp="2025-10-03 15:04:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:04:28.582368471 +0000 UTC m=+5637.785711888" watchObservedRunningTime="2025-10-03 15:04:28.587916556 +0000 UTC m=+5637.791259973" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.041519 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.044648 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.044695 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.120762 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.121094 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="dnsmasq-dns" containerID="cri-o://4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5" gracePeriod=10 Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.570987 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.618791 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb\") pod \"012bdeaa-9454-4893-996c-0d231f323c43\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.618849 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc\") pod \"012bdeaa-9454-4893-996c-0d231f323c43\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.618982 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb\") pod \"012bdeaa-9454-4893-996c-0d231f323c43\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.619001 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config\") pod \"012bdeaa-9454-4893-996c-0d231f323c43\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.619089 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bclkf\" (UniqueName: \"kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf\") pod \"012bdeaa-9454-4893-996c-0d231f323c43\" (UID: \"012bdeaa-9454-4893-996c-0d231f323c43\") " Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.629426 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf" (OuterVolumeSpecName: "kube-api-access-bclkf") pod "012bdeaa-9454-4893-996c-0d231f323c43" (UID: "012bdeaa-9454-4893-996c-0d231f323c43"). InnerVolumeSpecName "kube-api-access-bclkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.649764 4959 generic.go:334] "Generic (PLEG): container finished" podID="012bdeaa-9454-4893-996c-0d231f323c43" containerID="4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5" exitCode=0 Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.649807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" event={"ID":"012bdeaa-9454-4893-996c-0d231f323c43","Type":"ContainerDied","Data":"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5"} Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.649833 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" event={"ID":"012bdeaa-9454-4893-996c-0d231f323c43","Type":"ContainerDied","Data":"c257ad4b7f6b89786b152b12ec0d9e05b58fd4ca12f00afa149727fdc5940075"} Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.649848 4959 scope.go:117] "RemoveContainer" containerID="4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.649991 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596df78cd9-cwj4s" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.664520 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "012bdeaa-9454-4893-996c-0d231f323c43" (UID: "012bdeaa-9454-4893-996c-0d231f323c43"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.672924 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "012bdeaa-9454-4893-996c-0d231f323c43" (UID: "012bdeaa-9454-4893-996c-0d231f323c43"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.674616 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "012bdeaa-9454-4893-996c-0d231f323c43" (UID: "012bdeaa-9454-4893-996c-0d231f323c43"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.689555 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config" (OuterVolumeSpecName: "config") pod "012bdeaa-9454-4893-996c-0d231f323c43" (UID: "012bdeaa-9454-4893-996c-0d231f323c43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.708913 4959 scope.go:117] "RemoveContainer" containerID="56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.721129 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.721166 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.721177 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bclkf\" (UniqueName: \"kubernetes.io/projected/012bdeaa-9454-4893-996c-0d231f323c43-kube-api-access-bclkf\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.721202 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.721211 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/012bdeaa-9454-4893-996c-0d231f323c43-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.730670 4959 scope.go:117] "RemoveContainer" containerID="4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5" Oct 03 15:04:36 crc kubenswrapper[4959]: E1003 15:04:36.731436 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5\": container with ID starting with 4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5 not found: ID does not exist" containerID="4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.731484 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5"} err="failed to get container status \"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5\": rpc error: code = NotFound desc = could not find container \"4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5\": container with ID starting with 4ff8f84509f16169edf456e2a4e76ebbe66e48e1ee9ef3fa26b7e3a00af6cce5 not found: ID does not exist" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.731520 4959 scope.go:117] "RemoveContainer" containerID="56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d" Oct 03 15:04:36 crc kubenswrapper[4959]: E1003 15:04:36.731969 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d\": container with ID starting with 56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d not found: ID does not exist" containerID="56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.731991 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d"} err="failed to get container status \"56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d\": rpc error: code = NotFound desc = could not find container \"56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d\": container with ID starting with 56813c772d07002eee8c49e5a26d10990c1678417a66b9eab653a3514b15b68d not found: ID does not exist" Oct 03 15:04:36 crc kubenswrapper[4959]: I1003 15:04:36.994791 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:04:37 crc kubenswrapper[4959]: I1003 15:04:37.006006 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-596df78cd9-cwj4s"] Oct 03 15:04:37 crc kubenswrapper[4959]: I1003 15:04:37.697073 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="012bdeaa-9454-4893-996c-0d231f323c43" path="/var/lib/kubelet/pods/012bdeaa-9454-4893-996c-0d231f323c43/volumes" Oct 03 15:04:56 crc kubenswrapper[4959]: I1003 15:04:56.241890 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f668759cc-jmsxf" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.943737 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-mndcs"] Oct 03 15:05:03 crc kubenswrapper[4959]: E1003 15:05:03.944600 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="dnsmasq-dns" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.944615 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="dnsmasq-dns" Oct 03 15:05:03 crc kubenswrapper[4959]: E1003 15:05:03.944634 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="init" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.944640 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="init" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.944787 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="012bdeaa-9454-4893-996c-0d231f323c43" containerName="dnsmasq-dns" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.945318 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mndcs" Oct 03 15:05:03 crc kubenswrapper[4959]: I1003 15:05:03.953513 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mndcs"] Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.127100 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rkjp\" (UniqueName: \"kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp\") pod \"glance-db-create-mndcs\" (UID: \"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c\") " pod="openstack/glance-db-create-mndcs" Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.228878 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rkjp\" (UniqueName: \"kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp\") pod \"glance-db-create-mndcs\" (UID: \"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c\") " pod="openstack/glance-db-create-mndcs" Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.247586 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rkjp\" (UniqueName: \"kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp\") pod \"glance-db-create-mndcs\" (UID: \"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c\") " pod="openstack/glance-db-create-mndcs" Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.262561 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mndcs" Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.747981 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mndcs"] Oct 03 15:05:04 crc kubenswrapper[4959]: W1003 15:05:04.751332 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ea6cb3d_03f4_4eb2_8d94_ac9a9a8cf37c.slice/crio-bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a WatchSource:0}: Error finding container bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a: Status 404 returned error can't find the container with id bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a Oct 03 15:05:04 crc kubenswrapper[4959]: I1003 15:05:04.936142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mndcs" event={"ID":"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c","Type":"ContainerStarted","Data":"bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a"} Oct 03 15:05:05 crc kubenswrapper[4959]: I1003 15:05:05.827769 4959 scope.go:117] "RemoveContainer" containerID="2f8d800d6450c81e3af1f9ec6e909425c134f05dc3094f6f0ef288ed781edcf3" Oct 03 15:05:05 crc kubenswrapper[4959]: I1003 15:05:05.944677 4959 generic.go:334] "Generic (PLEG): container finished" podID="9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" containerID="cce3cda0e5531ca24833d5aed6642fc04c4a062b0358d1d4d0af87c6b35789be" exitCode=0 Oct 03 15:05:05 crc kubenswrapper[4959]: I1003 15:05:05.944718 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mndcs" event={"ID":"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c","Type":"ContainerDied","Data":"cce3cda0e5531ca24833d5aed6642fc04c4a062b0358d1d4d0af87c6b35789be"} Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.045025 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.045372 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.045498 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.046316 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.046519 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" gracePeriod=600 Oct 03 15:05:06 crc kubenswrapper[4959]: E1003 15:05:06.196949 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.961422 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" exitCode=0 Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.961547 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505"} Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.961638 4959 scope.go:117] "RemoveContainer" containerID="85a9693ed91fda4770b6f316a5388515fc1f0416f742fd2d96c2a1cf63c1ee9e" Oct 03 15:05:06 crc kubenswrapper[4959]: I1003 15:05:06.962504 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:05:06 crc kubenswrapper[4959]: E1003 15:05:06.963036 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.335727 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mndcs" Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.506025 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rkjp\" (UniqueName: \"kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp\") pod \"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c\" (UID: \"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c\") " Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.511663 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp" (OuterVolumeSpecName: "kube-api-access-4rkjp") pod "9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" (UID: "9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c"). InnerVolumeSpecName "kube-api-access-4rkjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.608845 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rkjp\" (UniqueName: \"kubernetes.io/projected/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c-kube-api-access-4rkjp\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.974424 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mndcs" Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.974363 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mndcs" event={"ID":"9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c","Type":"ContainerDied","Data":"bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a"} Oct 03 15:05:07 crc kubenswrapper[4959]: I1003 15:05:07.974689 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfadb14b5dc8453915858ec73b866feccc32c1123c3f9da3d21e939075b97a2a" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.987867 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-10f6-account-create-hmtcb"] Oct 03 15:05:13 crc kubenswrapper[4959]: E1003 15:05:13.988584 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" containerName="mariadb-database-create" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.988601 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" containerName="mariadb-database-create" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.988803 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" containerName="mariadb-database-create" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.989505 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.993055 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 03 15:05:13 crc kubenswrapper[4959]: I1003 15:05:13.997433 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-10f6-account-create-hmtcb"] Oct 03 15:05:14 crc kubenswrapper[4959]: I1003 15:05:14.121845 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7rpd\" (UniqueName: \"kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd\") pod \"glance-10f6-account-create-hmtcb\" (UID: \"2f542467-086c-4c9d-b634-d94240cdc81a\") " pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:14 crc kubenswrapper[4959]: I1003 15:05:14.223457 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7rpd\" (UniqueName: \"kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd\") pod \"glance-10f6-account-create-hmtcb\" (UID: \"2f542467-086c-4c9d-b634-d94240cdc81a\") " pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:14 crc kubenswrapper[4959]: I1003 15:05:14.244535 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7rpd\" (UniqueName: \"kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd\") pod \"glance-10f6-account-create-hmtcb\" (UID: \"2f542467-086c-4c9d-b634-d94240cdc81a\") " pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:14 crc kubenswrapper[4959]: I1003 15:05:14.309900 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:14 crc kubenswrapper[4959]: I1003 15:05:14.767504 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-10f6-account-create-hmtcb"] Oct 03 15:05:14 crc kubenswrapper[4959]: W1003 15:05:14.780057 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f542467_086c_4c9d_b634_d94240cdc81a.slice/crio-2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4 WatchSource:0}: Error finding container 2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4: Status 404 returned error can't find the container with id 2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4 Oct 03 15:05:15 crc kubenswrapper[4959]: I1003 15:05:15.049043 4959 generic.go:334] "Generic (PLEG): container finished" podID="2f542467-086c-4c9d-b634-d94240cdc81a" containerID="35a0a5560a92ef70bc58ba719dd4f61449e36a1224065ad4743e605d6b124fca" exitCode=0 Oct 03 15:05:15 crc kubenswrapper[4959]: I1003 15:05:15.049092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-10f6-account-create-hmtcb" event={"ID":"2f542467-086c-4c9d-b634-d94240cdc81a","Type":"ContainerDied","Data":"35a0a5560a92ef70bc58ba719dd4f61449e36a1224065ad4743e605d6b124fca"} Oct 03 15:05:15 crc kubenswrapper[4959]: I1003 15:05:15.049121 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-10f6-account-create-hmtcb" event={"ID":"2f542467-086c-4c9d-b634-d94240cdc81a","Type":"ContainerStarted","Data":"2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4"} Oct 03 15:05:16 crc kubenswrapper[4959]: I1003 15:05:16.354961 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:16 crc kubenswrapper[4959]: I1003 15:05:16.469021 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7rpd\" (UniqueName: \"kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd\") pod \"2f542467-086c-4c9d-b634-d94240cdc81a\" (UID: \"2f542467-086c-4c9d-b634-d94240cdc81a\") " Oct 03 15:05:16 crc kubenswrapper[4959]: I1003 15:05:16.474582 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd" (OuterVolumeSpecName: "kube-api-access-l7rpd") pod "2f542467-086c-4c9d-b634-d94240cdc81a" (UID: "2f542467-086c-4c9d-b634-d94240cdc81a"). InnerVolumeSpecName "kube-api-access-l7rpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:16 crc kubenswrapper[4959]: I1003 15:05:16.570842 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7rpd\" (UniqueName: \"kubernetes.io/projected/2f542467-086c-4c9d-b634-d94240cdc81a-kube-api-access-l7rpd\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:17 crc kubenswrapper[4959]: I1003 15:05:17.066953 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-10f6-account-create-hmtcb" event={"ID":"2f542467-086c-4c9d-b634-d94240cdc81a","Type":"ContainerDied","Data":"2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4"} Oct 03 15:05:17 crc kubenswrapper[4959]: I1003 15:05:17.066992 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fabf96a91f3ffc72fd653d592ad26f60129f27e76740abed4b336c732d474a4" Oct 03 15:05:17 crc kubenswrapper[4959]: I1003 15:05:17.067022 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-10f6-account-create-hmtcb" Oct 03 15:05:17 crc kubenswrapper[4959]: I1003 15:05:17.686905 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:05:17 crc kubenswrapper[4959]: E1003 15:05:17.687154 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.047704 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-klb8j"] Oct 03 15:05:19 crc kubenswrapper[4959]: E1003 15:05:19.048487 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f542467-086c-4c9d-b634-d94240cdc81a" containerName="mariadb-account-create" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.048508 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f542467-086c-4c9d-b634-d94240cdc81a" containerName="mariadb-account-create" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.048711 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f542467-086c-4c9d-b634-d94240cdc81a" containerName="mariadb-account-create" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.049445 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.052033 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.052738 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ctccn" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.064641 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-klb8j"] Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.223316 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.223590 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfdrw\" (UniqueName: \"kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.223694 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.223790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.324859 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.325181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.325308 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfdrw\" (UniqueName: \"kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.325413 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.333964 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.335016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.343986 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.349096 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfdrw\" (UniqueName: \"kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw\") pod \"glance-db-sync-klb8j\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.396394 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:19 crc kubenswrapper[4959]: I1003 15:05:19.956127 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-klb8j"] Oct 03 15:05:20 crc kubenswrapper[4959]: I1003 15:05:20.092097 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-klb8j" event={"ID":"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f","Type":"ContainerStarted","Data":"de51bb98d26a90601d72200c16698d8b0a57f5cc2ea7aacc39c936ae444c85ee"} Oct 03 15:05:21 crc kubenswrapper[4959]: I1003 15:05:21.104086 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-klb8j" event={"ID":"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f","Type":"ContainerStarted","Data":"061a55ca478075c7ca9d94bd6d87d1de94f390f1eade79ab396648f02a33b050"} Oct 03 15:05:21 crc kubenswrapper[4959]: I1003 15:05:21.130821 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-klb8j" podStartSLOduration=2.130801605 podStartE2EDuration="2.130801605s" podCreationTimestamp="2025-10-03 15:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:21.125096496 +0000 UTC m=+5690.328439923" watchObservedRunningTime="2025-10-03 15:05:21.130801605 +0000 UTC m=+5690.334145022" Oct 03 15:05:24 crc kubenswrapper[4959]: I1003 15:05:24.131476 4959 generic.go:334] "Generic (PLEG): container finished" podID="b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" containerID="061a55ca478075c7ca9d94bd6d87d1de94f390f1eade79ab396648f02a33b050" exitCode=0 Oct 03 15:05:24 crc kubenswrapper[4959]: I1003 15:05:24.131544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-klb8j" event={"ID":"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f","Type":"ContainerDied","Data":"061a55ca478075c7ca9d94bd6d87d1de94f390f1eade79ab396648f02a33b050"} Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.535206 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.637370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data\") pod \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.637428 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data\") pod \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.637606 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle\") pod \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.637682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfdrw\" (UniqueName: \"kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw\") pod \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\" (UID: \"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f\") " Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.644455 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" (UID: "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.644613 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw" (OuterVolumeSpecName: "kube-api-access-zfdrw") pod "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" (UID: "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f"). InnerVolumeSpecName "kube-api-access-zfdrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.666976 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" (UID: "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.695381 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data" (OuterVolumeSpecName: "config-data") pod "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" (UID: "b2c7fceb-49e7-47ca-940b-d2fc816c4f8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.740036 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.740277 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.740338 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:25 crc kubenswrapper[4959]: I1003 15:05:25.740394 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfdrw\" (UniqueName: \"kubernetes.io/projected/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f-kube-api-access-zfdrw\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.151066 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-klb8j" event={"ID":"b2c7fceb-49e7-47ca-940b-d2fc816c4f8f","Type":"ContainerDied","Data":"de51bb98d26a90601d72200c16698d8b0a57f5cc2ea7aacc39c936ae444c85ee"} Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.151472 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de51bb98d26a90601d72200c16698d8b0a57f5cc2ea7aacc39c936ae444c85ee" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.151117 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-klb8j" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.509124 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:26 crc kubenswrapper[4959]: E1003 15:05:26.509896 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" containerName="glance-db-sync" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.510045 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" containerName="glance-db-sync" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.510373 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" containerName="glance-db-sync" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.511541 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.515667 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.516797 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.516863 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.517078 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ctccn" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.521743 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.524423 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.526391 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.531723 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.588177 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.589572 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.593394 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.599903 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655093 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655473 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655516 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655554 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655577 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655593 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655609 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czdwq\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpmnk\" (UniqueName: \"kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.655692 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.756721 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wznpw\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.756778 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpmnk\" (UniqueName: \"kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.756829 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.756853 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.756914 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757611 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757646 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757688 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757713 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757811 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757885 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.757955 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758687 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758731 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758850 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758893 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758913 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758939 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.758956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czdwq\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.759286 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.759330 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.759527 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.759965 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.760695 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.762096 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.762770 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.762929 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.776949 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpmnk\" (UniqueName: \"kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk\") pod \"dnsmasq-dns-6b9b57f477-8r6wv\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.780157 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czdwq\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq\") pod \"glance-default-external-api-0\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.844426 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.854169 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.860229 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wznpw\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.860536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.860747 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.860985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.861161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.861376 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.861561 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.862280 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.862799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.865034 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.865078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.865957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.879305 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.885217 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wznpw\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw\") pod \"glance-default-internal-api-0\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:26 crc kubenswrapper[4959]: I1003 15:05:26.919343 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:27 crc kubenswrapper[4959]: I1003 15:05:27.540265 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.005181 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.185806 4959 generic.go:334] "Generic (PLEG): container finished" podID="120521cb-93fa-4296-91d0-36c110b8a972" containerID="7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5" exitCode=0 Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.185874 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" event={"ID":"120521cb-93fa-4296-91d0-36c110b8a972","Type":"ContainerDied","Data":"7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5"} Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.185927 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" event={"ID":"120521cb-93fa-4296-91d0-36c110b8a972","Type":"ContainerStarted","Data":"fadb30857003f4fa90600b995ea1ce19b0e7a4e5fa28d1ea124cd7c1efafeb11"} Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.325505 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:28 crc kubenswrapper[4959]: I1003 15:05:28.685864 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:05:28 crc kubenswrapper[4959]: E1003 15:05:28.686259 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.195865 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerStarted","Data":"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c"} Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.196156 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerStarted","Data":"748a780056cfe2b9c9da52965033db3ff6b0ffd9efb85bc73c5f30e2fb4000cb"} Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.199414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" event={"ID":"120521cb-93fa-4296-91d0-36c110b8a972","Type":"ContainerStarted","Data":"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a"} Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.200296 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.221161 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" podStartSLOduration=3.221147489 podStartE2EDuration="3.221147489s" podCreationTimestamp="2025-10-03 15:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:29.218969876 +0000 UTC m=+5698.422313293" watchObservedRunningTime="2025-10-03 15:05:29.221147489 +0000 UTC m=+5698.424490906" Oct 03 15:05:29 crc kubenswrapper[4959]: I1003 15:05:29.418174 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:29 crc kubenswrapper[4959]: W1003 15:05:29.427760 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b118a43_56d6_4e5f_9880_5f8a8c87832b.slice/crio-97fffacdd05be22f64cfb8f8d4384639e6d37c341d588e4da6a5050f443b9de4 WatchSource:0}: Error finding container 97fffacdd05be22f64cfb8f8d4384639e6d37c341d588e4da6a5050f443b9de4: Status 404 returned error can't find the container with id 97fffacdd05be22f64cfb8f8d4384639e6d37c341d588e4da6a5050f443b9de4 Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.214637 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerStarted","Data":"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418"} Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.214792 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-httpd" containerID="cri-o://b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" gracePeriod=30 Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.215114 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-log" containerID="cri-o://56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" gracePeriod=30 Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.217825 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerStarted","Data":"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581"} Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.217895 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerStarted","Data":"97fffacdd05be22f64cfb8f8d4384639e6d37c341d588e4da6a5050f443b9de4"} Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.237044 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.237025127 podStartE2EDuration="4.237025127s" podCreationTimestamp="2025-10-03 15:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:30.232952357 +0000 UTC m=+5699.436295814" watchObservedRunningTime="2025-10-03 15:05:30.237025127 +0000 UTC m=+5699.440368544" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.310540 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.786405 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.945964 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946106 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czdwq\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946169 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946244 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946287 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946310 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946326 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run\") pod \"86d69241-c3e3-4709-b840-31391fe562da\" (UID: \"86d69241-c3e3-4709-b840-31391fe562da\") " Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946518 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs" (OuterVolumeSpecName: "logs") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.946722 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.950424 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.957310 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq" (OuterVolumeSpecName: "kube-api-access-czdwq") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "kube-api-access-czdwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.966400 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph" (OuterVolumeSpecName: "ceph") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:30 crc kubenswrapper[4959]: I1003 15:05:30.973359 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts" (OuterVolumeSpecName: "scripts") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.002315 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.031959 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data" (OuterVolumeSpecName: "config-data") pod "86d69241-c3e3-4709-b840-31391fe562da" (UID: "86d69241-c3e3-4709-b840-31391fe562da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048647 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czdwq\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-kube-api-access-czdwq\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048683 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048693 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86d69241-c3e3-4709-b840-31391fe562da-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048703 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048714 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d69241-c3e3-4709-b840-31391fe562da-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.048723 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86d69241-c3e3-4709-b840-31391fe562da-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.229337 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerStarted","Data":"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252"} Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.230016 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-httpd" containerID="cri-o://9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" gracePeriod=30 Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.230086 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-log" containerID="cri-o://48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" gracePeriod=30 Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.237766 4959 generic.go:334] "Generic (PLEG): container finished" podID="86d69241-c3e3-4709-b840-31391fe562da" containerID="b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" exitCode=0 Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.238097 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.238100 4959 generic.go:334] "Generic (PLEG): container finished" podID="86d69241-c3e3-4709-b840-31391fe562da" containerID="56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" exitCode=143 Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.237947 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerDied","Data":"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418"} Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.238440 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerDied","Data":"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c"} Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.238496 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86d69241-c3e3-4709-b840-31391fe562da","Type":"ContainerDied","Data":"748a780056cfe2b9c9da52965033db3ff6b0ffd9efb85bc73c5f30e2fb4000cb"} Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.238538 4959 scope.go:117] "RemoveContainer" containerID="b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.269367 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.269331996 podStartE2EDuration="5.269331996s" podCreationTimestamp="2025-10-03 15:05:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:31.257732033 +0000 UTC m=+5700.461075490" watchObservedRunningTime="2025-10-03 15:05:31.269331996 +0000 UTC m=+5700.472675483" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.285420 4959 scope.go:117] "RemoveContainer" containerID="56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.301433 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.313399 4959 scope.go:117] "RemoveContainer" containerID="b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" Oct 03 15:05:31 crc kubenswrapper[4959]: E1003 15:05:31.315161 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418\": container with ID starting with b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418 not found: ID does not exist" containerID="b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.315279 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418"} err="failed to get container status \"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418\": rpc error: code = NotFound desc = could not find container \"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418\": container with ID starting with b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418 not found: ID does not exist" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.315320 4959 scope.go:117] "RemoveContainer" containerID="56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" Oct 03 15:05:31 crc kubenswrapper[4959]: E1003 15:05:31.315774 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c\": container with ID starting with 56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c not found: ID does not exist" containerID="56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.315815 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c"} err="failed to get container status \"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c\": rpc error: code = NotFound desc = could not find container \"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c\": container with ID starting with 56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c not found: ID does not exist" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.315839 4959 scope.go:117] "RemoveContainer" containerID="b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.316066 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418"} err="failed to get container status \"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418\": rpc error: code = NotFound desc = could not find container \"b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418\": container with ID starting with b3f427740b86ffb6251940f755766e9e77ecae82167f2866e7b130afb9c0c418 not found: ID does not exist" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.316099 4959 scope.go:117] "RemoveContainer" containerID="56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.316348 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c"} err="failed to get container status \"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c\": rpc error: code = NotFound desc = could not find container \"56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c\": container with ID starting with 56dacc0a2cbc2a16f52a542da1d9aadd5cea53fd04feb7f4d793556f51e5c15c not found: ID does not exist" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.316642 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.328932 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:31 crc kubenswrapper[4959]: E1003 15:05:31.329618 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-httpd" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.329740 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-httpd" Oct 03 15:05:31 crc kubenswrapper[4959]: E1003 15:05:31.329913 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-log" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.330008 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-log" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.330378 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-log" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.330992 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d69241-c3e3-4709-b840-31391fe562da" containerName="glance-httpd" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.332691 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.343425 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.355961 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374081 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7flv\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374136 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374449 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374511 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374606 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.374702 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.476500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.476721 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.476876 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7flv\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.477601 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.478098 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.478666 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.479082 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.478605 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.479609 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.483799 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.484642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.484816 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.488075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.494784 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7flv\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv\") pod \"glance-default-external-api-0\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " pod="openstack/glance-default-external-api-0" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.700437 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d69241-c3e3-4709-b840-31391fe562da" path="/var/lib/kubelet/pods/86d69241-c3e3-4709-b840-31391fe562da/volumes" Oct 03 15:05:31 crc kubenswrapper[4959]: I1003 15:05:31.737728 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.014936 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094332 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094425 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094515 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094550 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094565 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094579 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094604 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wznpw\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw\") pod \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\" (UID: \"4b118a43-56d6-4e5f-9880-5f8a8c87832b\") " Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.094878 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.095543 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs" (OuterVolumeSpecName: "logs") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.101441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw" (OuterVolumeSpecName: "kube-api-access-wznpw") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "kube-api-access-wznpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.101767 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph" (OuterVolumeSpecName: "ceph") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.107567 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts" (OuterVolumeSpecName: "scripts") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.120727 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.135787 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data" (OuterVolumeSpecName: "config-data") pod "4b118a43-56d6-4e5f-9880-5f8a8c87832b" (UID: "4b118a43-56d6-4e5f-9880-5f8a8c87832b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196905 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196937 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196946 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b118a43-56d6-4e5f-9880-5f8a8c87832b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196956 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196965 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wznpw\" (UniqueName: \"kubernetes.io/projected/4b118a43-56d6-4e5f-9880-5f8a8c87832b-kube-api-access-wznpw\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196977 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.196985 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b118a43-56d6-4e5f-9880-5f8a8c87832b-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252449 4959 generic.go:334] "Generic (PLEG): container finished" podID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerID="9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" exitCode=0 Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252477 4959 generic.go:334] "Generic (PLEG): container finished" podID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerID="48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" exitCode=143 Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252496 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerDied","Data":"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252"} Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252521 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerDied","Data":"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581"} Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252526 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252544 4959 scope.go:117] "RemoveContainer" containerID="9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.252533 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4b118a43-56d6-4e5f-9880-5f8a8c87832b","Type":"ContainerDied","Data":"97fffacdd05be22f64cfb8f8d4384639e6d37c341d588e4da6a5050f443b9de4"} Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.282095 4959 scope.go:117] "RemoveContainer" containerID="48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.303838 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.311221 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.317970 4959 scope.go:117] "RemoveContainer" containerID="9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" Oct 03 15:05:32 crc kubenswrapper[4959]: E1003 15:05:32.318498 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252\": container with ID starting with 9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252 not found: ID does not exist" containerID="9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.318525 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252"} err="failed to get container status \"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252\": rpc error: code = NotFound desc = could not find container \"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252\": container with ID starting with 9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252 not found: ID does not exist" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.318545 4959 scope.go:117] "RemoveContainer" containerID="48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" Oct 03 15:05:32 crc kubenswrapper[4959]: E1003 15:05:32.318778 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581\": container with ID starting with 48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581 not found: ID does not exist" containerID="48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.318812 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581"} err="failed to get container status \"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581\": rpc error: code = NotFound desc = could not find container \"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581\": container with ID starting with 48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581 not found: ID does not exist" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.318836 4959 scope.go:117] "RemoveContainer" containerID="9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.319051 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252"} err="failed to get container status \"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252\": rpc error: code = NotFound desc = could not find container \"9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252\": container with ID starting with 9a25d651b44d4db121ab7934a5066b95b08e03a8734f69b8eb4ccac710294252 not found: ID does not exist" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.319065 4959 scope.go:117] "RemoveContainer" containerID="48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.319248 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581"} err="failed to get container status \"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581\": rpc error: code = NotFound desc = could not find container \"48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581\": container with ID starting with 48eb32d2efb8462175cbdbfaa22035b4ec24d9128c776460e12b3a1abc87f581 not found: ID does not exist" Oct 03 15:05:32 crc kubenswrapper[4959]: W1003 15:05:32.322861 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc546427c_b9ad_4355_867a_7b90a42bd5c9.slice/crio-e02881ce0c0240589163b758f7727fce9b1c1195e47f4e223023444ad383be28 WatchSource:0}: Error finding container e02881ce0c0240589163b758f7727fce9b1c1195e47f4e223023444ad383be28: Status 404 returned error can't find the container with id e02881ce0c0240589163b758f7727fce9b1c1195e47f4e223023444ad383be28 Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.333994 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.338762 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:32 crc kubenswrapper[4959]: E1003 15:05:32.339299 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-httpd" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.339324 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-httpd" Oct 03 15:05:32 crc kubenswrapper[4959]: E1003 15:05:32.339344 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-log" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.339353 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-log" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.339547 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-httpd" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.339570 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" containerName="glance-log" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.340546 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.345701 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.346953 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.400652 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.400799 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9khr\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.400975 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.401050 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.401407 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.401492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.401527 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.503848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.503915 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9khr\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.503981 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.504002 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.504177 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.504234 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.504255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.509074 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.509126 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.509474 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.509489 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.512358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.513276 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.550112 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9khr\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr\") pod \"glance-default-internal-api-0\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:05:32 crc kubenswrapper[4959]: I1003 15:05:32.670534 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:33 crc kubenswrapper[4959]: I1003 15:05:33.181163 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:05:33 crc kubenswrapper[4959]: I1003 15:05:33.272658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerStarted","Data":"b803fd51295e215eae9411738e10fe8923b5148585e6b75c37e71b66d6a252d4"} Oct 03 15:05:33 crc kubenswrapper[4959]: I1003 15:05:33.273023 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerStarted","Data":"e02881ce0c0240589163b758f7727fce9b1c1195e47f4e223023444ad383be28"} Oct 03 15:05:33 crc kubenswrapper[4959]: I1003 15:05:33.277779 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerStarted","Data":"f49facb801a5ac64a308caeaeaccc019ae8f51b268c1e034579edf7ee72352cb"} Oct 03 15:05:33 crc kubenswrapper[4959]: I1003 15:05:33.726095 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b118a43-56d6-4e5f-9880-5f8a8c87832b" path="/var/lib/kubelet/pods/4b118a43-56d6-4e5f-9880-5f8a8c87832b/volumes" Oct 03 15:05:34 crc kubenswrapper[4959]: I1003 15:05:34.288925 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerStarted","Data":"2bbff60e17c67bb687d16fe1789ba3aaa49f65fdfd66759a47f6769d92a146b9"} Oct 03 15:05:34 crc kubenswrapper[4959]: I1003 15:05:34.289304 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerStarted","Data":"b50cf27ea1530345a2dfa9e9f67ae26f85c66ad093aebeefc2e5dc15adb45c0c"} Oct 03 15:05:34 crc kubenswrapper[4959]: I1003 15:05:34.291885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerStarted","Data":"43f8f5ac309a5bb879d4e686502c58a47a358376d14d8d63685b476f5fa43540"} Oct 03 15:05:34 crc kubenswrapper[4959]: I1003 15:05:34.312229 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.312208402 podStartE2EDuration="2.312208402s" podCreationTimestamp="2025-10-03 15:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:34.307531548 +0000 UTC m=+5703.510874975" watchObservedRunningTime="2025-10-03 15:05:34.312208402 +0000 UTC m=+5703.515551819" Oct 03 15:05:34 crc kubenswrapper[4959]: I1003 15:05:34.336372 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.336351733 podStartE2EDuration="3.336351733s" podCreationTimestamp="2025-10-03 15:05:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:05:34.332450688 +0000 UTC m=+5703.535794125" watchObservedRunningTime="2025-10-03 15:05:34.336351733 +0000 UTC m=+5703.539695150" Oct 03 15:05:36 crc kubenswrapper[4959]: I1003 15:05:36.856445 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:05:36 crc kubenswrapper[4959]: I1003 15:05:36.931133 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:05:36 crc kubenswrapper[4959]: I1003 15:05:36.931552 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="dnsmasq-dns" containerID="cri-o://f4225ff6cb567f3b52fcf43640e9f093f08c003aa49a53a289bd097d6f67bd09" gracePeriod=10 Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.319860 4959 generic.go:334] "Generic (PLEG): container finished" podID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerID="f4225ff6cb567f3b52fcf43640e9f093f08c003aa49a53a289bd097d6f67bd09" exitCode=0 Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.319894 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" event={"ID":"44e8681b-60fc-48b1-b9b1-80e36865fde0","Type":"ContainerDied","Data":"f4225ff6cb567f3b52fcf43640e9f093f08c003aa49a53a289bd097d6f67bd09"} Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.404265 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.500870 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb\") pod \"44e8681b-60fc-48b1-b9b1-80e36865fde0\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.501223 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb\") pod \"44e8681b-60fc-48b1-b9b1-80e36865fde0\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.501267 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config\") pod \"44e8681b-60fc-48b1-b9b1-80e36865fde0\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.501370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc\") pod \"44e8681b-60fc-48b1-b9b1-80e36865fde0\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.501427 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4hkz\" (UniqueName: \"kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz\") pod \"44e8681b-60fc-48b1-b9b1-80e36865fde0\" (UID: \"44e8681b-60fc-48b1-b9b1-80e36865fde0\") " Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.507831 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz" (OuterVolumeSpecName: "kube-api-access-v4hkz") pod "44e8681b-60fc-48b1-b9b1-80e36865fde0" (UID: "44e8681b-60fc-48b1-b9b1-80e36865fde0"). InnerVolumeSpecName "kube-api-access-v4hkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.584259 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config" (OuterVolumeSpecName: "config") pod "44e8681b-60fc-48b1-b9b1-80e36865fde0" (UID: "44e8681b-60fc-48b1-b9b1-80e36865fde0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.587983 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "44e8681b-60fc-48b1-b9b1-80e36865fde0" (UID: "44e8681b-60fc-48b1-b9b1-80e36865fde0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.589800 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "44e8681b-60fc-48b1-b9b1-80e36865fde0" (UID: "44e8681b-60fc-48b1-b9b1-80e36865fde0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.589936 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "44e8681b-60fc-48b1-b9b1-80e36865fde0" (UID: "44e8681b-60fc-48b1-b9b1-80e36865fde0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.603795 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.603835 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4hkz\" (UniqueName: \"kubernetes.io/projected/44e8681b-60fc-48b1-b9b1-80e36865fde0-kube-api-access-v4hkz\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.603847 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.603878 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:37 crc kubenswrapper[4959]: I1003 15:05:37.603889 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e8681b-60fc-48b1-b9b1-80e36865fde0-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.351621 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" event={"ID":"44e8681b-60fc-48b1-b9b1-80e36865fde0","Type":"ContainerDied","Data":"c51b62d539fcb3526cb2da20ab2c86c3c4339e530f2f5dbc568fe981a48b2f8f"} Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.351685 4959 scope.go:117] "RemoveContainer" containerID="f4225ff6cb567f3b52fcf43640e9f093f08c003aa49a53a289bd097d6f67bd09" Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.351859 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59d59797-wbthw" Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.378561 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.384831 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59d59797-wbthw"] Oct 03 15:05:38 crc kubenswrapper[4959]: I1003 15:05:38.403578 4959 scope.go:117] "RemoveContainer" containerID="c145715c157804ce34ff13794d9e659ea101155b8c5ff957ea75ef56d4465bba" Oct 03 15:05:39 crc kubenswrapper[4959]: I1003 15:05:39.694583 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" path="/var/lib/kubelet/pods/44e8681b-60fc-48b1-b9b1-80e36865fde0/volumes" Oct 03 15:05:40 crc kubenswrapper[4959]: I1003 15:05:40.686800 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:05:40 crc kubenswrapper[4959]: E1003 15:05:40.688063 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:41 crc kubenswrapper[4959]: I1003 15:05:41.738564 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 15:05:41 crc kubenswrapper[4959]: I1003 15:05:41.738650 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 15:05:41 crc kubenswrapper[4959]: I1003 15:05:41.786868 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 15:05:41 crc kubenswrapper[4959]: I1003 15:05:41.816286 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.394358 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.394405 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.672809 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.672932 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.709112 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:42 crc kubenswrapper[4959]: I1003 15:05:42.720564 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:43 crc kubenswrapper[4959]: I1003 15:05:43.402695 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:43 crc kubenswrapper[4959]: I1003 15:05:43.402741 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:44 crc kubenswrapper[4959]: I1003 15:05:44.456327 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 15:05:44 crc kubenswrapper[4959]: I1003 15:05:44.456457 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 15:05:44 crc kubenswrapper[4959]: I1003 15:05:44.609793 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 15:05:45 crc kubenswrapper[4959]: I1003 15:05:45.514804 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:45 crc kubenswrapper[4959]: I1003 15:05:45.515142 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 15:05:45 crc kubenswrapper[4959]: I1003 15:05:45.521887 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.826584 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vdxhq"] Oct 03 15:05:51 crc kubenswrapper[4959]: E1003 15:05:51.829437 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="init" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.829609 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="init" Oct 03 15:05:51 crc kubenswrapper[4959]: E1003 15:05:51.829749 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="dnsmasq-dns" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.829870 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="dnsmasq-dns" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.830326 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="44e8681b-60fc-48b1-b9b1-80e36865fde0" containerName="dnsmasq-dns" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.831513 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.847458 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vdxhq"] Oct 03 15:05:51 crc kubenswrapper[4959]: I1003 15:05:51.965266 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxr4p\" (UniqueName: \"kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p\") pod \"placement-db-create-vdxhq\" (UID: \"e5135668-739c-49a6-a7e5-645acd5e8021\") " pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:52 crc kubenswrapper[4959]: I1003 15:05:52.067710 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxr4p\" (UniqueName: \"kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p\") pod \"placement-db-create-vdxhq\" (UID: \"e5135668-739c-49a6-a7e5-645acd5e8021\") " pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:52 crc kubenswrapper[4959]: I1003 15:05:52.085728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxr4p\" (UniqueName: \"kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p\") pod \"placement-db-create-vdxhq\" (UID: \"e5135668-739c-49a6-a7e5-645acd5e8021\") " pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:52 crc kubenswrapper[4959]: I1003 15:05:52.157555 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:52 crc kubenswrapper[4959]: I1003 15:05:52.593502 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vdxhq"] Oct 03 15:05:52 crc kubenswrapper[4959]: W1003 15:05:52.596078 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5135668_739c_49a6_a7e5_645acd5e8021.slice/crio-4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa WatchSource:0}: Error finding container 4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa: Status 404 returned error can't find the container with id 4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa Oct 03 15:05:52 crc kubenswrapper[4959]: I1003 15:05:52.686806 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:05:52 crc kubenswrapper[4959]: E1003 15:05:52.687450 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:05:53 crc kubenswrapper[4959]: I1003 15:05:53.487699 4959 generic.go:334] "Generic (PLEG): container finished" podID="e5135668-739c-49a6-a7e5-645acd5e8021" containerID="f5cae178a0cb97b7ab5fb961af6ffe5aad0714fb7fca08b2e365cb2dc15ac427" exitCode=0 Oct 03 15:05:53 crc kubenswrapper[4959]: I1003 15:05:53.487740 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vdxhq" event={"ID":"e5135668-739c-49a6-a7e5-645acd5e8021","Type":"ContainerDied","Data":"f5cae178a0cb97b7ab5fb961af6ffe5aad0714fb7fca08b2e365cb2dc15ac427"} Oct 03 15:05:53 crc kubenswrapper[4959]: I1003 15:05:53.487808 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vdxhq" event={"ID":"e5135668-739c-49a6-a7e5-645acd5e8021","Type":"ContainerStarted","Data":"4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa"} Oct 03 15:05:54 crc kubenswrapper[4959]: I1003 15:05:54.814471 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vdxhq" Oct 03 15:05:54 crc kubenswrapper[4959]: I1003 15:05:54.816314 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxr4p\" (UniqueName: \"kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p\") pod \"e5135668-739c-49a6-a7e5-645acd5e8021\" (UID: \"e5135668-739c-49a6-a7e5-645acd5e8021\") " Oct 03 15:05:54 crc kubenswrapper[4959]: I1003 15:05:54.821586 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p" (OuterVolumeSpecName: "kube-api-access-dxr4p") pod "e5135668-739c-49a6-a7e5-645acd5e8021" (UID: "e5135668-739c-49a6-a7e5-645acd5e8021"). InnerVolumeSpecName "kube-api-access-dxr4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:05:54 crc kubenswrapper[4959]: I1003 15:05:54.917881 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxr4p\" (UniqueName: \"kubernetes.io/projected/e5135668-739c-49a6-a7e5-645acd5e8021-kube-api-access-dxr4p\") on node \"crc\" DevicePath \"\"" Oct 03 15:05:55 crc kubenswrapper[4959]: I1003 15:05:55.509697 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vdxhq" event={"ID":"e5135668-739c-49a6-a7e5-645acd5e8021","Type":"ContainerDied","Data":"4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa"} Oct 03 15:05:55 crc kubenswrapper[4959]: I1003 15:05:55.509747 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a02b942fa499ba3330e2461784de8ca8097c73ab975d5c310e8735ef5613ffa" Oct 03 15:05:55 crc kubenswrapper[4959]: I1003 15:05:55.509787 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vdxhq" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.973030 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-569f-account-create-j5b44"] Oct 03 15:06:01 crc kubenswrapper[4959]: E1003 15:06:01.974453 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5135668-739c-49a6-a7e5-645acd5e8021" containerName="mariadb-database-create" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.974477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5135668-739c-49a6-a7e5-645acd5e8021" containerName="mariadb-database-create" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.974780 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5135668-739c-49a6-a7e5-645acd5e8021" containerName="mariadb-database-create" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.975924 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.978397 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 03 15:06:01 crc kubenswrapper[4959]: I1003 15:06:01.981682 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569f-account-create-j5b44"] Oct 03 15:06:02 crc kubenswrapper[4959]: I1003 15:06:02.162699 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79lc\" (UniqueName: \"kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc\") pod \"placement-569f-account-create-j5b44\" (UID: \"46315488-a177-4eb1-ab05-e91501e1b11b\") " pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:02 crc kubenswrapper[4959]: I1003 15:06:02.264185 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79lc\" (UniqueName: \"kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc\") pod \"placement-569f-account-create-j5b44\" (UID: \"46315488-a177-4eb1-ab05-e91501e1b11b\") " pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:02 crc kubenswrapper[4959]: I1003 15:06:02.289727 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79lc\" (UniqueName: \"kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc\") pod \"placement-569f-account-create-j5b44\" (UID: \"46315488-a177-4eb1-ab05-e91501e1b11b\") " pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:02 crc kubenswrapper[4959]: I1003 15:06:02.300868 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:02 crc kubenswrapper[4959]: I1003 15:06:02.738561 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-569f-account-create-j5b44"] Oct 03 15:06:03 crc kubenswrapper[4959]: I1003 15:06:03.586949 4959 generic.go:334] "Generic (PLEG): container finished" podID="46315488-a177-4eb1-ab05-e91501e1b11b" containerID="0ee5999a6cd6d503eadb5ed5c19b9fea27fea8409722de7618e2fdd61e9683e6" exitCode=0 Oct 03 15:06:03 crc kubenswrapper[4959]: I1003 15:06:03.587034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569f-account-create-j5b44" event={"ID":"46315488-a177-4eb1-ab05-e91501e1b11b","Type":"ContainerDied","Data":"0ee5999a6cd6d503eadb5ed5c19b9fea27fea8409722de7618e2fdd61e9683e6"} Oct 03 15:06:03 crc kubenswrapper[4959]: I1003 15:06:03.587472 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569f-account-create-j5b44" event={"ID":"46315488-a177-4eb1-ab05-e91501e1b11b","Type":"ContainerStarted","Data":"0250ca9e9260827c8e6dced341d0d503b375b51d23fa43cef6351a411100df8e"} Oct 03 15:06:04 crc kubenswrapper[4959]: I1003 15:06:04.685081 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:06:04 crc kubenswrapper[4959]: E1003 15:06:04.685511 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:06:04 crc kubenswrapper[4959]: I1003 15:06:04.927491 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.114277 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h79lc\" (UniqueName: \"kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc\") pod \"46315488-a177-4eb1-ab05-e91501e1b11b\" (UID: \"46315488-a177-4eb1-ab05-e91501e1b11b\") " Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.123910 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc" (OuterVolumeSpecName: "kube-api-access-h79lc") pod "46315488-a177-4eb1-ab05-e91501e1b11b" (UID: "46315488-a177-4eb1-ab05-e91501e1b11b"). InnerVolumeSpecName "kube-api-access-h79lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.216108 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h79lc\" (UniqueName: \"kubernetes.io/projected/46315488-a177-4eb1-ab05-e91501e1b11b-kube-api-access-h79lc\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.625264 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-569f-account-create-j5b44" event={"ID":"46315488-a177-4eb1-ab05-e91501e1b11b","Type":"ContainerDied","Data":"0250ca9e9260827c8e6dced341d0d503b375b51d23fa43cef6351a411100df8e"} Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.625305 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0250ca9e9260827c8e6dced341d0d503b375b51d23fa43cef6351a411100df8e" Oct 03 15:06:05 crc kubenswrapper[4959]: I1003 15:06:05.625369 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-569f-account-create-j5b44" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.266751 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:06:07 crc kubenswrapper[4959]: E1003 15:06:07.267487 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46315488-a177-4eb1-ab05-e91501e1b11b" containerName="mariadb-account-create" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.267507 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="46315488-a177-4eb1-ab05-e91501e1b11b" containerName="mariadb-account-create" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.267725 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="46315488-a177-4eb1-ab05-e91501e1b11b" containerName="mariadb-account-create" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.268875 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.299674 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.332432 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-c6pxg"] Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.333522 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.337837 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.338064 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-smrtp" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.338231 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.341389 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-c6pxg"] Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.358505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.358786 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.358913 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359347 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359387 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359442 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz5rk\" (UniqueName: \"kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359655 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzv9n\" (UniqueName: \"kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359714 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.359878 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462028 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzv9n\" (UniqueName: \"kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462097 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462121 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462155 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462241 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462284 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462321 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.462435 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz5rk\" (UniqueName: \"kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.464176 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.464800 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.464828 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.464987 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.466152 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.478510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.478827 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.479793 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.481776 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz5rk\" (UniqueName: \"kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk\") pod \"placement-db-sync-c6pxg\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.483012 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzv9n\" (UniqueName: \"kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n\") pod \"dnsmasq-dns-74df65d56c-6ch7g\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.595160 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:07 crc kubenswrapper[4959]: I1003 15:06:07.655480 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.044528 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:06:08 crc kubenswrapper[4959]: W1003 15:06:08.048674 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3d8c4df_a879_4a83_9f3e_8b1a4b6d0f0c.slice/crio-865ea30b7bb3afdca7a74fe26a3a403c8dfb839490767f741d8f60787bc222c2 WatchSource:0}: Error finding container 865ea30b7bb3afdca7a74fe26a3a403c8dfb839490767f741d8f60787bc222c2: Status 404 returned error can't find the container with id 865ea30b7bb3afdca7a74fe26a3a403c8dfb839490767f741d8f60787bc222c2 Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.147973 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-c6pxg"] Oct 03 15:06:08 crc kubenswrapper[4959]: W1003 15:06:08.151102 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d19694_00e8_4ad4_93b9_e016048c49bf.slice/crio-df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a WatchSource:0}: Error finding container df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a: Status 404 returned error can't find the container with id df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.649624 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c6pxg" event={"ID":"27d19694-00e8-4ad4-93b9-e016048c49bf","Type":"ContainerStarted","Data":"e096bcae9c5073352458aef380dde247924485f3c52922c9d2d4d2f9255bc6d6"} Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.649988 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c6pxg" event={"ID":"27d19694-00e8-4ad4-93b9-e016048c49bf","Type":"ContainerStarted","Data":"df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a"} Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.651503 4959 generic.go:334] "Generic (PLEG): container finished" podID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerID="ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93" exitCode=0 Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.651554 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" event={"ID":"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c","Type":"ContainerDied","Data":"ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93"} Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.651586 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" event={"ID":"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c","Type":"ContainerStarted","Data":"865ea30b7bb3afdca7a74fe26a3a403c8dfb839490767f741d8f60787bc222c2"} Oct 03 15:06:08 crc kubenswrapper[4959]: I1003 15:06:08.691000 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-c6pxg" podStartSLOduration=1.690980851 podStartE2EDuration="1.690980851s" podCreationTimestamp="2025-10-03 15:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:06:08.685896706 +0000 UTC m=+5737.889240123" watchObservedRunningTime="2025-10-03 15:06:08.690980851 +0000 UTC m=+5737.894324288" Oct 03 15:06:09 crc kubenswrapper[4959]: I1003 15:06:09.661011 4959 generic.go:334] "Generic (PLEG): container finished" podID="27d19694-00e8-4ad4-93b9-e016048c49bf" containerID="e096bcae9c5073352458aef380dde247924485f3c52922c9d2d4d2f9255bc6d6" exitCode=0 Oct 03 15:06:09 crc kubenswrapper[4959]: I1003 15:06:09.661081 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c6pxg" event={"ID":"27d19694-00e8-4ad4-93b9-e016048c49bf","Type":"ContainerDied","Data":"e096bcae9c5073352458aef380dde247924485f3c52922c9d2d4d2f9255bc6d6"} Oct 03 15:06:09 crc kubenswrapper[4959]: I1003 15:06:09.664237 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" event={"ID":"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c","Type":"ContainerStarted","Data":"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8"} Oct 03 15:06:09 crc kubenswrapper[4959]: I1003 15:06:09.664600 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:09 crc kubenswrapper[4959]: I1003 15:06:09.703941 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" podStartSLOduration=2.703922396 podStartE2EDuration="2.703922396s" podCreationTimestamp="2025-10-03 15:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:06:09.701725803 +0000 UTC m=+5738.905069230" watchObservedRunningTime="2025-10-03 15:06:09.703922396 +0000 UTC m=+5738.907265813" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.032321 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.132820 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data\") pod \"27d19694-00e8-4ad4-93b9-e016048c49bf\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.132886 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts\") pod \"27d19694-00e8-4ad4-93b9-e016048c49bf\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.132968 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle\") pod \"27d19694-00e8-4ad4-93b9-e016048c49bf\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.133004 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs\") pod \"27d19694-00e8-4ad4-93b9-e016048c49bf\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.133082 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz5rk\" (UniqueName: \"kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk\") pod \"27d19694-00e8-4ad4-93b9-e016048c49bf\" (UID: \"27d19694-00e8-4ad4-93b9-e016048c49bf\") " Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.133592 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs" (OuterVolumeSpecName: "logs") pod "27d19694-00e8-4ad4-93b9-e016048c49bf" (UID: "27d19694-00e8-4ad4-93b9-e016048c49bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.138762 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts" (OuterVolumeSpecName: "scripts") pod "27d19694-00e8-4ad4-93b9-e016048c49bf" (UID: "27d19694-00e8-4ad4-93b9-e016048c49bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.139347 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk" (OuterVolumeSpecName: "kube-api-access-sz5rk") pod "27d19694-00e8-4ad4-93b9-e016048c49bf" (UID: "27d19694-00e8-4ad4-93b9-e016048c49bf"). InnerVolumeSpecName "kube-api-access-sz5rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.160916 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data" (OuterVolumeSpecName: "config-data") pod "27d19694-00e8-4ad4-93b9-e016048c49bf" (UID: "27d19694-00e8-4ad4-93b9-e016048c49bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.161009 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27d19694-00e8-4ad4-93b9-e016048c49bf" (UID: "27d19694-00e8-4ad4-93b9-e016048c49bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.235667 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.235714 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d19694-00e8-4ad4-93b9-e016048c49bf-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.235727 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz5rk\" (UniqueName: \"kubernetes.io/projected/27d19694-00e8-4ad4-93b9-e016048c49bf-kube-api-access-sz5rk\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.235743 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.235754 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d19694-00e8-4ad4-93b9-e016048c49bf-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.692018 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-c6pxg" Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.708221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-c6pxg" event={"ID":"27d19694-00e8-4ad4-93b9-e016048c49bf","Type":"ContainerDied","Data":"df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a"} Oct 03 15:06:11 crc kubenswrapper[4959]: I1003 15:06:11.708284 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df191b3470a5fbe6ece169d6a289c55d8b8740381c92933e1bdac84367f9eb0a" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.128252 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-544547cb8d-86sh8"] Oct 03 15:06:12 crc kubenswrapper[4959]: E1003 15:06:12.128617 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d19694-00e8-4ad4-93b9-e016048c49bf" containerName="placement-db-sync" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.128630 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d19694-00e8-4ad4-93b9-e016048c49bf" containerName="placement-db-sync" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.128802 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d19694-00e8-4ad4-93b9-e016048c49bf" containerName="placement-db-sync" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.129740 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.135507 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-smrtp" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.135731 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.135833 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.143014 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-544547cb8d-86sh8"] Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.254183 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff4d98d-ab89-42e4-b79c-c7088858d882-logs\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.254912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-combined-ca-bundle\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.254962 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-scripts\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.255045 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4dvv\" (UniqueName: \"kubernetes.io/projected/bff4d98d-ab89-42e4-b79c-c7088858d882-kube-api-access-r4dvv\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.255077 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-config-data\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.356498 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-combined-ca-bundle\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.356559 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-scripts\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.356630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4dvv\" (UniqueName: \"kubernetes.io/projected/bff4d98d-ab89-42e4-b79c-c7088858d882-kube-api-access-r4dvv\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.356654 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-config-data\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.356704 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff4d98d-ab89-42e4-b79c-c7088858d882-logs\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.357164 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff4d98d-ab89-42e4-b79c-c7088858d882-logs\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.362395 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-scripts\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.362855 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-config-data\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.363364 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff4d98d-ab89-42e4-b79c-c7088858d882-combined-ca-bundle\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.373054 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4dvv\" (UniqueName: \"kubernetes.io/projected/bff4d98d-ab89-42e4-b79c-c7088858d882-kube-api-access-r4dvv\") pod \"placement-544547cb8d-86sh8\" (UID: \"bff4d98d-ab89-42e4-b79c-c7088858d882\") " pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.448964 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:12 crc kubenswrapper[4959]: I1003 15:06:12.930448 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-544547cb8d-86sh8"] Oct 03 15:06:12 crc kubenswrapper[4959]: W1003 15:06:12.938391 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbff4d98d_ab89_42e4_b79c_c7088858d882.slice/crio-f84faccf55569a19708eb5fd7225f4869e570351f1c8ed7d42676e34f0b96b8f WatchSource:0}: Error finding container f84faccf55569a19708eb5fd7225f4869e570351f1c8ed7d42676e34f0b96b8f: Status 404 returned error can't find the container with id f84faccf55569a19708eb5fd7225f4869e570351f1c8ed7d42676e34f0b96b8f Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.715639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-544547cb8d-86sh8" event={"ID":"bff4d98d-ab89-42e4-b79c-c7088858d882","Type":"ContainerStarted","Data":"ebb04a442129827f675d0754bb9fe1b9224dd274cc2cfbc0aa02c3ea8951ec5c"} Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.716067 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.716086 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.716097 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-544547cb8d-86sh8" event={"ID":"bff4d98d-ab89-42e4-b79c-c7088858d882","Type":"ContainerStarted","Data":"615f8fb1c5cc0cc31cc1e68a02d2e6c76f139e4f12066f94729e9eb24e4268f7"} Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.716110 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-544547cb8d-86sh8" event={"ID":"bff4d98d-ab89-42e4-b79c-c7088858d882","Type":"ContainerStarted","Data":"f84faccf55569a19708eb5fd7225f4869e570351f1c8ed7d42676e34f0b96b8f"} Oct 03 15:06:13 crc kubenswrapper[4959]: I1003 15:06:13.744430 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-544547cb8d-86sh8" podStartSLOduration=1.744405774 podStartE2EDuration="1.744405774s" podCreationTimestamp="2025-10-03 15:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:06:13.737627008 +0000 UTC m=+5742.940970435" watchObservedRunningTime="2025-10-03 15:06:13.744405774 +0000 UTC m=+5742.947749201" Oct 03 15:06:17 crc kubenswrapper[4959]: I1003 15:06:17.597173 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:06:17 crc kubenswrapper[4959]: I1003 15:06:17.674569 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:06:17 crc kubenswrapper[4959]: I1003 15:06:17.674778 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="dnsmasq-dns" containerID="cri-o://e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a" gracePeriod=10 Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.198642 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.283060 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config\") pod \"120521cb-93fa-4296-91d0-36c110b8a972\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.283146 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb\") pod \"120521cb-93fa-4296-91d0-36c110b8a972\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.283181 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc\") pod \"120521cb-93fa-4296-91d0-36c110b8a972\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.283273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpmnk\" (UniqueName: \"kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk\") pod \"120521cb-93fa-4296-91d0-36c110b8a972\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.283354 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb\") pod \"120521cb-93fa-4296-91d0-36c110b8a972\" (UID: \"120521cb-93fa-4296-91d0-36c110b8a972\") " Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.289099 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk" (OuterVolumeSpecName: "kube-api-access-zpmnk") pod "120521cb-93fa-4296-91d0-36c110b8a972" (UID: "120521cb-93fa-4296-91d0-36c110b8a972"). InnerVolumeSpecName "kube-api-access-zpmnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.328612 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "120521cb-93fa-4296-91d0-36c110b8a972" (UID: "120521cb-93fa-4296-91d0-36c110b8a972"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.331825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config" (OuterVolumeSpecName: "config") pod "120521cb-93fa-4296-91d0-36c110b8a972" (UID: "120521cb-93fa-4296-91d0-36c110b8a972"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.336403 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "120521cb-93fa-4296-91d0-36c110b8a972" (UID: "120521cb-93fa-4296-91d0-36c110b8a972"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.341806 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "120521cb-93fa-4296-91d0-36c110b8a972" (UID: "120521cb-93fa-4296-91d0-36c110b8a972"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.384920 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.385167 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.385176 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.385185 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpmnk\" (UniqueName: \"kubernetes.io/projected/120521cb-93fa-4296-91d0-36c110b8a972-kube-api-access-zpmnk\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.385208 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/120521cb-93fa-4296-91d0-36c110b8a972-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.763211 4959 generic.go:334] "Generic (PLEG): container finished" podID="120521cb-93fa-4296-91d0-36c110b8a972" containerID="e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a" exitCode=0 Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.763263 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" event={"ID":"120521cb-93fa-4296-91d0-36c110b8a972","Type":"ContainerDied","Data":"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a"} Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.763301 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" event={"ID":"120521cb-93fa-4296-91d0-36c110b8a972","Type":"ContainerDied","Data":"fadb30857003f4fa90600b995ea1ce19b0e7a4e5fa28d1ea124cd7c1efafeb11"} Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.763317 4959 scope.go:117] "RemoveContainer" containerID="e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.765877 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b9b57f477-8r6wv" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.797042 4959 scope.go:117] "RemoveContainer" containerID="7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.814711 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.821526 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b9b57f477-8r6wv"] Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.843035 4959 scope.go:117] "RemoveContainer" containerID="e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a" Oct 03 15:06:18 crc kubenswrapper[4959]: E1003 15:06:18.843614 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a\": container with ID starting with e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a not found: ID does not exist" containerID="e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.843660 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a"} err="failed to get container status \"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a\": rpc error: code = NotFound desc = could not find container \"e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a\": container with ID starting with e8e9e17e058cc09c8ce9f138cf56076271d57cd32fa0adc97641158b11c7af1a not found: ID does not exist" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.843695 4959 scope.go:117] "RemoveContainer" containerID="7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5" Oct 03 15:06:18 crc kubenswrapper[4959]: E1003 15:06:18.844099 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5\": container with ID starting with 7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5 not found: ID does not exist" containerID="7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5" Oct 03 15:06:18 crc kubenswrapper[4959]: I1003 15:06:18.844125 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5"} err="failed to get container status \"7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5\": rpc error: code = NotFound desc = could not find container \"7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5\": container with ID starting with 7e72e3ea841f0a03cb745b5a228b3a7698b17bbf442d2b6d0826f78356f255a5 not found: ID does not exist" Oct 03 15:06:19 crc kubenswrapper[4959]: I1003 15:06:19.686140 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:06:19 crc kubenswrapper[4959]: E1003 15:06:19.686625 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:06:19 crc kubenswrapper[4959]: I1003 15:06:19.700550 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="120521cb-93fa-4296-91d0-36c110b8a972" path="/var/lib/kubelet/pods/120521cb-93fa-4296-91d0-36c110b8a972/volumes" Oct 03 15:06:30 crc kubenswrapper[4959]: I1003 15:06:30.685400 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:06:30 crc kubenswrapper[4959]: E1003 15:06:30.686136 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:06:41 crc kubenswrapper[4959]: I1003 15:06:41.700521 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:06:41 crc kubenswrapper[4959]: E1003 15:06:41.701482 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:06:43 crc kubenswrapper[4959]: I1003 15:06:43.499120 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:43 crc kubenswrapper[4959]: I1003 15:06:43.641091 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-544547cb8d-86sh8" Oct 03 15:06:55 crc kubenswrapper[4959]: I1003 15:06:55.687244 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:06:55 crc kubenswrapper[4959]: E1003 15:06:55.687883 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.726114 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-m2q8h"] Oct 03 15:07:04 crc kubenswrapper[4959]: E1003 15:07:04.727061 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="init" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.727077 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="init" Oct 03 15:07:04 crc kubenswrapper[4959]: E1003 15:07:04.727098 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="dnsmasq-dns" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.727106 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="dnsmasq-dns" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.727405 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="120521cb-93fa-4296-91d0-36c110b8a972" containerName="dnsmasq-dns" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.728129 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.734425 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m2q8h"] Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.750733 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtv2r\" (UniqueName: \"kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r\") pod \"nova-api-db-create-m2q8h\" (UID: \"5015457d-5742-4f2e-82b1-e9da287f6332\") " pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.809930 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-69rpg"] Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.811018 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.826965 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-69rpg"] Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.853014 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrk66\" (UniqueName: \"kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66\") pod \"nova-cell0-db-create-69rpg\" (UID: \"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5\") " pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.853420 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtv2r\" (UniqueName: \"kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r\") pod \"nova-api-db-create-m2q8h\" (UID: \"5015457d-5742-4f2e-82b1-e9da287f6332\") " pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.870474 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtv2r\" (UniqueName: \"kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r\") pod \"nova-api-db-create-m2q8h\" (UID: \"5015457d-5742-4f2e-82b1-e9da287f6332\") " pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.921611 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-qdks2"] Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.922774 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.935290 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qdks2"] Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.973291 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrk66\" (UniqueName: \"kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66\") pod \"nova-cell0-db-create-69rpg\" (UID: \"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5\") " pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.973388 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2nqz\" (UniqueName: \"kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz\") pod \"nova-cell1-db-create-qdks2\" (UID: \"a79b8d66-b0a9-4a18-ae85-574106748f8a\") " pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:04 crc kubenswrapper[4959]: I1003 15:07:04.993504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrk66\" (UniqueName: \"kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66\") pod \"nova-cell0-db-create-69rpg\" (UID: \"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5\") " pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.047118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.075261 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2nqz\" (UniqueName: \"kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz\") pod \"nova-cell1-db-create-qdks2\" (UID: \"a79b8d66-b0a9-4a18-ae85-574106748f8a\") " pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.095341 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2nqz\" (UniqueName: \"kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz\") pod \"nova-cell1-db-create-qdks2\" (UID: \"a79b8d66-b0a9-4a18-ae85-574106748f8a\") " pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.127636 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.282171 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.477719 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-m2q8h"] Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.572771 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-69rpg"] Oct 03 15:07:05 crc kubenswrapper[4959]: W1003 15:07:05.590085 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d4256b_825f_4ee1_b1de_3d8db5cf6ed5.slice/crio-73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861 WatchSource:0}: Error finding container 73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861: Status 404 returned error can't find the container with id 73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861 Oct 03 15:07:05 crc kubenswrapper[4959]: I1003 15:07:05.716974 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qdks2"] Oct 03 15:07:05 crc kubenswrapper[4959]: W1003 15:07:05.777432 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda79b8d66_b0a9_4a18_ae85_574106748f8a.slice/crio-1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c WatchSource:0}: Error finding container 1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c: Status 404 returned error can't find the container with id 1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.014089 4959 scope.go:117] "RemoveContainer" containerID="539f03434b7f8a428d16b72e99611072319094bf66e0c78eea29090d45daa4b8" Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.245534 4959 generic.go:334] "Generic (PLEG): container finished" podID="45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" containerID="0345bfd3ca139d031001ec4291e82685a5d5f5c411f3e8c535f85bce80318bfe" exitCode=0 Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.245599 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69rpg" event={"ID":"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5","Type":"ContainerDied","Data":"0345bfd3ca139d031001ec4291e82685a5d5f5c411f3e8c535f85bce80318bfe"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.245625 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69rpg" event={"ID":"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5","Type":"ContainerStarted","Data":"73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.247443 4959 generic.go:334] "Generic (PLEG): container finished" podID="5015457d-5742-4f2e-82b1-e9da287f6332" containerID="3eba02b4c8ab1dda9828a6b93401119c54a34d1ccb3ffb22475357bc91df9312" exitCode=0 Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.247483 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m2q8h" event={"ID":"5015457d-5742-4f2e-82b1-e9da287f6332","Type":"ContainerDied","Data":"3eba02b4c8ab1dda9828a6b93401119c54a34d1ccb3ffb22475357bc91df9312"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.247500 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m2q8h" event={"ID":"5015457d-5742-4f2e-82b1-e9da287f6332","Type":"ContainerStarted","Data":"c759fc445c2ce01b86b02e8b1a4c0683cf25e23c08733d675fb9c641db358c85"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.249101 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdks2" event={"ID":"a79b8d66-b0a9-4a18-ae85-574106748f8a","Type":"ContainerStarted","Data":"c3a5a97c89d7dbf3c92010ae6b739394db37ec5dfe1f4b205d4d3f0a4397277e"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.249130 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdks2" event={"ID":"a79b8d66-b0a9-4a18-ae85-574106748f8a","Type":"ContainerStarted","Data":"1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c"} Oct 03 15:07:06 crc kubenswrapper[4959]: I1003 15:07:06.314384 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-qdks2" podStartSLOduration=2.314368853 podStartE2EDuration="2.314368853s" podCreationTimestamp="2025-10-03 15:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:06.310649392 +0000 UTC m=+5795.513992809" watchObservedRunningTime="2025-10-03 15:07:06.314368853 +0000 UTC m=+5795.517712270" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.261273 4959 generic.go:334] "Generic (PLEG): container finished" podID="a79b8d66-b0a9-4a18-ae85-574106748f8a" containerID="c3a5a97c89d7dbf3c92010ae6b739394db37ec5dfe1f4b205d4d3f0a4397277e" exitCode=0 Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.261366 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdks2" event={"ID":"a79b8d66-b0a9-4a18-ae85-574106748f8a","Type":"ContainerDied","Data":"c3a5a97c89d7dbf3c92010ae6b739394db37ec5dfe1f4b205d4d3f0a4397277e"} Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.689745 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.695496 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.840880 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtv2r\" (UniqueName: \"kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r\") pod \"5015457d-5742-4f2e-82b1-e9da287f6332\" (UID: \"5015457d-5742-4f2e-82b1-e9da287f6332\") " Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.841074 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrk66\" (UniqueName: \"kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66\") pod \"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5\" (UID: \"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5\") " Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.846222 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66" (OuterVolumeSpecName: "kube-api-access-mrk66") pod "45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" (UID: "45d4256b-825f-4ee1-b1de-3d8db5cf6ed5"). InnerVolumeSpecName "kube-api-access-mrk66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.848269 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r" (OuterVolumeSpecName: "kube-api-access-vtv2r") pod "5015457d-5742-4f2e-82b1-e9da287f6332" (UID: "5015457d-5742-4f2e-82b1-e9da287f6332"). InnerVolumeSpecName "kube-api-access-vtv2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.942929 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrk66\" (UniqueName: \"kubernetes.io/projected/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5-kube-api-access-mrk66\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:07 crc kubenswrapper[4959]: I1003 15:07:07.942958 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtv2r\" (UniqueName: \"kubernetes.io/projected/5015457d-5742-4f2e-82b1-e9da287f6332-kube-api-access-vtv2r\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.275372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-69rpg" event={"ID":"45d4256b-825f-4ee1-b1de-3d8db5cf6ed5","Type":"ContainerDied","Data":"73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861"} Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.275416 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73cdacacfb51c0c51f522a6a3eda50f6e863bb32981a990b6b8aba24eb3f6861" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.275447 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-69rpg" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.277976 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-m2q8h" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.277993 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-m2q8h" event={"ID":"5015457d-5742-4f2e-82b1-e9da287f6332","Type":"ContainerDied","Data":"c759fc445c2ce01b86b02e8b1a4c0683cf25e23c08733d675fb9c641db358c85"} Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.278049 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c759fc445c2ce01b86b02e8b1a4c0683cf25e23c08733d675fb9c641db358c85" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.715324 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.761656 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2nqz\" (UniqueName: \"kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz\") pod \"a79b8d66-b0a9-4a18-ae85-574106748f8a\" (UID: \"a79b8d66-b0a9-4a18-ae85-574106748f8a\") " Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.764590 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz" (OuterVolumeSpecName: "kube-api-access-d2nqz") pod "a79b8d66-b0a9-4a18-ae85-574106748f8a" (UID: "a79b8d66-b0a9-4a18-ae85-574106748f8a"). InnerVolumeSpecName "kube-api-access-d2nqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:08 crc kubenswrapper[4959]: I1003 15:07:08.863927 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2nqz\" (UniqueName: \"kubernetes.io/projected/a79b8d66-b0a9-4a18-ae85-574106748f8a-kube-api-access-d2nqz\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:09 crc kubenswrapper[4959]: I1003 15:07:09.289033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdks2" event={"ID":"a79b8d66-b0a9-4a18-ae85-574106748f8a","Type":"ContainerDied","Data":"1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c"} Oct 03 15:07:09 crc kubenswrapper[4959]: I1003 15:07:09.289073 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1229b9a803af0dd9415bf84079b49366eda256762a7395ca53b3a651d1bea31c" Oct 03 15:07:09 crc kubenswrapper[4959]: I1003 15:07:09.289127 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdks2" Oct 03 15:07:09 crc kubenswrapper[4959]: I1003 15:07:09.685823 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:07:09 crc kubenswrapper[4959]: E1003 15:07:09.686104 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.873597 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db25-account-create-tccj6"] Oct 03 15:07:14 crc kubenswrapper[4959]: E1003 15:07:14.874757 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.874779 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: E1003 15:07:14.874804 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5015457d-5742-4f2e-82b1-e9da287f6332" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.874816 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5015457d-5742-4f2e-82b1-e9da287f6332" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: E1003 15:07:14.874900 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79b8d66-b0a9-4a18-ae85-574106748f8a" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.874913 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79b8d66-b0a9-4a18-ae85-574106748f8a" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.875259 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79b8d66-b0a9-4a18-ae85-574106748f8a" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.875306 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5015457d-5742-4f2e-82b1-e9da287f6332" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.875335 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" containerName="mariadb-database-create" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.876404 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.879685 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.898818 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db25-account-create-tccj6"] Oct 03 15:07:14 crc kubenswrapper[4959]: I1003 15:07:14.982694 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9c6p\" (UniqueName: \"kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p\") pod \"nova-api-db25-account-create-tccj6\" (UID: \"7ac5a899-a2cd-4d75-85a9-fdab2926e695\") " pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.056555 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-3366-account-create-h2hwt"] Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.058031 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.060966 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.066423 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3366-account-create-h2hwt"] Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.084411 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9c6p\" (UniqueName: \"kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p\") pod \"nova-api-db25-account-create-tccj6\" (UID: \"7ac5a899-a2cd-4d75-85a9-fdab2926e695\") " pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.103124 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9c6p\" (UniqueName: \"kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p\") pod \"nova-api-db25-account-create-tccj6\" (UID: \"7ac5a899-a2cd-4d75-85a9-fdab2926e695\") " pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.186820 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrjkt\" (UniqueName: \"kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt\") pod \"nova-cell0-3366-account-create-h2hwt\" (UID: \"f3dbff33-0ddf-4691-ae33-4190fa73a648\") " pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.199949 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.257754 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b5b8-account-create-4qrx2"] Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.259209 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.261764 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.269938 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b5b8-account-create-4qrx2"] Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.288994 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrjkt\" (UniqueName: \"kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt\") pod \"nova-cell0-3366-account-create-h2hwt\" (UID: \"f3dbff33-0ddf-4691-ae33-4190fa73a648\") " pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.309635 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrjkt\" (UniqueName: \"kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt\") pod \"nova-cell0-3366-account-create-h2hwt\" (UID: \"f3dbff33-0ddf-4691-ae33-4190fa73a648\") " pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.373847 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.391914 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm6f7\" (UniqueName: \"kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7\") pod \"nova-cell1-b5b8-account-create-4qrx2\" (UID: \"35a92e66-c252-4c32-9f83-05d555300a64\") " pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.493831 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm6f7\" (UniqueName: \"kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7\") pod \"nova-cell1-b5b8-account-create-4qrx2\" (UID: \"35a92e66-c252-4c32-9f83-05d555300a64\") " pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.510681 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm6f7\" (UniqueName: \"kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7\") pod \"nova-cell1-b5b8-account-create-4qrx2\" (UID: \"35a92e66-c252-4c32-9f83-05d555300a64\") " pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.660413 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.664848 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db25-account-create-tccj6"] Oct 03 15:07:15 crc kubenswrapper[4959]: W1003 15:07:15.666036 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac5a899_a2cd_4d75_85a9_fdab2926e695.slice/crio-8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c WatchSource:0}: Error finding container 8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c: Status 404 returned error can't find the container with id 8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c Oct 03 15:07:15 crc kubenswrapper[4959]: I1003 15:07:15.839927 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-3366-account-create-h2hwt"] Oct 03 15:07:15 crc kubenswrapper[4959]: W1003 15:07:15.844367 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3dbff33_0ddf_4691_ae33_4190fa73a648.slice/crio-ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0 WatchSource:0}: Error finding container ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0: Status 404 returned error can't find the container with id ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0 Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.072713 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b5b8-account-create-4qrx2"] Oct 03 15:07:16 crc kubenswrapper[4959]: W1003 15:07:16.074546 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35a92e66_c252_4c32_9f83_05d555300a64.slice/crio-30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b WatchSource:0}: Error finding container 30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b: Status 404 returned error can't find the container with id 30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.377926 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" event={"ID":"35a92e66-c252-4c32-9f83-05d555300a64","Type":"ContainerStarted","Data":"30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b"} Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.396736 4959 generic.go:334] "Generic (PLEG): container finished" podID="7ac5a899-a2cd-4d75-85a9-fdab2926e695" containerID="6dd46a12ddbf3dba9ccebc3d17f454910073bad46a8a2e4082d2c192b7402413" exitCode=0 Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.396857 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db25-account-create-tccj6" event={"ID":"7ac5a899-a2cd-4d75-85a9-fdab2926e695","Type":"ContainerDied","Data":"6dd46a12ddbf3dba9ccebc3d17f454910073bad46a8a2e4082d2c192b7402413"} Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.396898 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db25-account-create-tccj6" event={"ID":"7ac5a899-a2cd-4d75-85a9-fdab2926e695","Type":"ContainerStarted","Data":"8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c"} Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.400029 4959 generic.go:334] "Generic (PLEG): container finished" podID="f3dbff33-0ddf-4691-ae33-4190fa73a648" containerID="e8c96f23ea891ef8cc63d4f472fbf52821a7c50318ee5ed0071763b50e185bf6" exitCode=0 Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.400099 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3366-account-create-h2hwt" event={"ID":"f3dbff33-0ddf-4691-ae33-4190fa73a648","Type":"ContainerDied","Data":"e8c96f23ea891ef8cc63d4f472fbf52821a7c50318ee5ed0071763b50e185bf6"} Oct 03 15:07:16 crc kubenswrapper[4959]: I1003 15:07:16.400131 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3366-account-create-h2hwt" event={"ID":"f3dbff33-0ddf-4691-ae33-4190fa73a648","Type":"ContainerStarted","Data":"ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0"} Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.418089 4959 generic.go:334] "Generic (PLEG): container finished" podID="35a92e66-c252-4c32-9f83-05d555300a64" containerID="21511c2b47aa72040fb41b48089d01a0aaab3c2ac55eff25fb5742ef7e6e051f" exitCode=0 Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.418618 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" event={"ID":"35a92e66-c252-4c32-9f83-05d555300a64","Type":"ContainerDied","Data":"21511c2b47aa72040fb41b48089d01a0aaab3c2ac55eff25fb5742ef7e6e051f"} Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.866039 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.873356 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.956166 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrjkt\" (UniqueName: \"kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt\") pod \"f3dbff33-0ddf-4691-ae33-4190fa73a648\" (UID: \"f3dbff33-0ddf-4691-ae33-4190fa73a648\") " Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.956350 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9c6p\" (UniqueName: \"kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p\") pod \"7ac5a899-a2cd-4d75-85a9-fdab2926e695\" (UID: \"7ac5a899-a2cd-4d75-85a9-fdab2926e695\") " Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.961745 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt" (OuterVolumeSpecName: "kube-api-access-mrjkt") pod "f3dbff33-0ddf-4691-ae33-4190fa73a648" (UID: "f3dbff33-0ddf-4691-ae33-4190fa73a648"). InnerVolumeSpecName "kube-api-access-mrjkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:17 crc kubenswrapper[4959]: I1003 15:07:17.962061 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p" (OuterVolumeSpecName: "kube-api-access-h9c6p") pod "7ac5a899-a2cd-4d75-85a9-fdab2926e695" (UID: "7ac5a899-a2cd-4d75-85a9-fdab2926e695"). InnerVolumeSpecName "kube-api-access-h9c6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.058177 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrjkt\" (UniqueName: \"kubernetes.io/projected/f3dbff33-0ddf-4691-ae33-4190fa73a648-kube-api-access-mrjkt\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.058273 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9c6p\" (UniqueName: \"kubernetes.io/projected/7ac5a899-a2cd-4d75-85a9-fdab2926e695-kube-api-access-h9c6p\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.432687 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-3366-account-create-h2hwt" event={"ID":"f3dbff33-0ddf-4691-ae33-4190fa73a648","Type":"ContainerDied","Data":"ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0"} Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.432746 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec4119cd85eb15232b573e7c2dbb0a8a5f423c0ed1ccf0b061b57b8cc20dabd0" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.432829 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-3366-account-create-h2hwt" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.436368 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db25-account-create-tccj6" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.436378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db25-account-create-tccj6" event={"ID":"7ac5a899-a2cd-4d75-85a9-fdab2926e695","Type":"ContainerDied","Data":"8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c"} Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.436437 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc3811fa1062d69732962696c32d9c1702aa2775eeef0f089802c9853679f6c" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.691026 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.790078 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm6f7\" (UniqueName: \"kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7\") pod \"35a92e66-c252-4c32-9f83-05d555300a64\" (UID: \"35a92e66-c252-4c32-9f83-05d555300a64\") " Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.793337 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7" (OuterVolumeSpecName: "kube-api-access-tm6f7") pod "35a92e66-c252-4c32-9f83-05d555300a64" (UID: "35a92e66-c252-4c32-9f83-05d555300a64"). InnerVolumeSpecName "kube-api-access-tm6f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:18 crc kubenswrapper[4959]: I1003 15:07:18.892224 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm6f7\" (UniqueName: \"kubernetes.io/projected/35a92e66-c252-4c32-9f83-05d555300a64-kube-api-access-tm6f7\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:19 crc kubenswrapper[4959]: I1003 15:07:19.446968 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" event={"ID":"35a92e66-c252-4c32-9f83-05d555300a64","Type":"ContainerDied","Data":"30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b"} Oct 03 15:07:19 crc kubenswrapper[4959]: I1003 15:07:19.447009 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30907d786b2d82440fb484f852357c3cb808eda4c01d43f05df9616070ceb46b" Oct 03 15:07:19 crc kubenswrapper[4959]: I1003 15:07:19.447015 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b5b8-account-create-4qrx2" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.299323 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tbfhj"] Oct 03 15:07:20 crc kubenswrapper[4959]: E1003 15:07:20.300264 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3dbff33-0ddf-4691-ae33-4190fa73a648" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300344 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3dbff33-0ddf-4691-ae33-4190fa73a648" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: E1003 15:07:20.300423 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a92e66-c252-4c32-9f83-05d555300a64" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300486 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a92e66-c252-4c32-9f83-05d555300a64" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: E1003 15:07:20.300548 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac5a899-a2cd-4d75-85a9-fdab2926e695" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300619 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac5a899-a2cd-4d75-85a9-fdab2926e695" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300831 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3dbff33-0ddf-4691-ae33-4190fa73a648" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300901 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a92e66-c252-4c32-9f83-05d555300a64" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.300954 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac5a899-a2cd-4d75-85a9-fdab2926e695" containerName="mariadb-account-create" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.301591 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.303915 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9v64s" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.305155 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.305454 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.320692 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tbfhj"] Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.419766 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.419844 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.419868 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.419921 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xk29\" (UniqueName: \"kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.522050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.523389 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.523599 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.523862 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xk29\" (UniqueName: \"kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.528379 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.529343 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.530071 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.542471 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xk29\" (UniqueName: \"kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29\") pod \"nova-cell0-conductor-db-sync-tbfhj\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.622491 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:20 crc kubenswrapper[4959]: I1003 15:07:20.884062 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tbfhj"] Oct 03 15:07:21 crc kubenswrapper[4959]: I1003 15:07:21.465775 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" event={"ID":"a035a96b-0609-470c-9bf1-066f2b76387c","Type":"ContainerStarted","Data":"4a8746be50e608988b75314ad682cf7cbcadb2f4a3f65b66161430079f22f794"} Oct 03 15:07:21 crc kubenswrapper[4959]: I1003 15:07:21.465817 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" event={"ID":"a035a96b-0609-470c-9bf1-066f2b76387c","Type":"ContainerStarted","Data":"d012e14618a54d2f08f499d30c679982a74fcbdfa447191615f9f50dd53af046"} Oct 03 15:07:21 crc kubenswrapper[4959]: I1003 15:07:21.486992 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" podStartSLOduration=1.486975132 podStartE2EDuration="1.486975132s" podCreationTimestamp="2025-10-03 15:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:21.481033847 +0000 UTC m=+5810.684377264" watchObservedRunningTime="2025-10-03 15:07:21.486975132 +0000 UTC m=+5810.690318549" Oct 03 15:07:21 crc kubenswrapper[4959]: I1003 15:07:21.701658 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:07:21 crc kubenswrapper[4959]: E1003 15:07:21.702089 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:07:26 crc kubenswrapper[4959]: I1003 15:07:26.513897 4959 generic.go:334] "Generic (PLEG): container finished" podID="a035a96b-0609-470c-9bf1-066f2b76387c" containerID="4a8746be50e608988b75314ad682cf7cbcadb2f4a3f65b66161430079f22f794" exitCode=0 Oct 03 15:07:26 crc kubenswrapper[4959]: I1003 15:07:26.513997 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" event={"ID":"a035a96b-0609-470c-9bf1-066f2b76387c","Type":"ContainerDied","Data":"4a8746be50e608988b75314ad682cf7cbcadb2f4a3f65b66161430079f22f794"} Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.919375 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.965461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts\") pod \"a035a96b-0609-470c-9bf1-066f2b76387c\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.965524 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data\") pod \"a035a96b-0609-470c-9bf1-066f2b76387c\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.965557 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle\") pod \"a035a96b-0609-470c-9bf1-066f2b76387c\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.965581 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xk29\" (UniqueName: \"kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29\") pod \"a035a96b-0609-470c-9bf1-066f2b76387c\" (UID: \"a035a96b-0609-470c-9bf1-066f2b76387c\") " Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.970742 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29" (OuterVolumeSpecName: "kube-api-access-8xk29") pod "a035a96b-0609-470c-9bf1-066f2b76387c" (UID: "a035a96b-0609-470c-9bf1-066f2b76387c"). InnerVolumeSpecName "kube-api-access-8xk29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.972080 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts" (OuterVolumeSpecName: "scripts") pod "a035a96b-0609-470c-9bf1-066f2b76387c" (UID: "a035a96b-0609-470c-9bf1-066f2b76387c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:27 crc kubenswrapper[4959]: I1003 15:07:27.992372 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data" (OuterVolumeSpecName: "config-data") pod "a035a96b-0609-470c-9bf1-066f2b76387c" (UID: "a035a96b-0609-470c-9bf1-066f2b76387c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.011972 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a035a96b-0609-470c-9bf1-066f2b76387c" (UID: "a035a96b-0609-470c-9bf1-066f2b76387c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.067337 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.067379 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.067391 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a035a96b-0609-470c-9bf1-066f2b76387c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.067404 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xk29\" (UniqueName: \"kubernetes.io/projected/a035a96b-0609-470c-9bf1-066f2b76387c-kube-api-access-8xk29\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.534354 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" event={"ID":"a035a96b-0609-470c-9bf1-066f2b76387c","Type":"ContainerDied","Data":"d012e14618a54d2f08f499d30c679982a74fcbdfa447191615f9f50dd53af046"} Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.534735 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d012e14618a54d2f08f499d30c679982a74fcbdfa447191615f9f50dd53af046" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.534428 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tbfhj" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.612836 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:07:28 crc kubenswrapper[4959]: E1003 15:07:28.613562 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a035a96b-0609-470c-9bf1-066f2b76387c" containerName="nova-cell0-conductor-db-sync" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.613738 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a035a96b-0609-470c-9bf1-066f2b76387c" containerName="nova-cell0-conductor-db-sync" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.614073 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a035a96b-0609-470c-9bf1-066f2b76387c" containerName="nova-cell0-conductor-db-sync" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.614917 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.618040 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.618260 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9v64s" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.635255 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.677769 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.677865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppwwz\" (UniqueName: \"kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.677909 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.780036 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.780138 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppwwz\" (UniqueName: \"kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.780174 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.786987 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.792077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.798625 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppwwz\" (UniqueName: \"kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz\") pod \"nova-cell0-conductor-0\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:28 crc kubenswrapper[4959]: I1003 15:07:28.945592 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:29 crc kubenswrapper[4959]: I1003 15:07:29.422245 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:07:29 crc kubenswrapper[4959]: I1003 15:07:29.547654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4d460ff8-f592-4b23-9650-e0f08fa3daa3","Type":"ContainerStarted","Data":"2edea1e3bcfdb47e75f51085e8af5efc901a0fb7612e3e6189d54984ca92ea80"} Oct 03 15:07:30 crc kubenswrapper[4959]: I1003 15:07:30.559383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4d460ff8-f592-4b23-9650-e0f08fa3daa3","Type":"ContainerStarted","Data":"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f"} Oct 03 15:07:30 crc kubenswrapper[4959]: I1003 15:07:30.559701 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:30 crc kubenswrapper[4959]: I1003 15:07:30.587189 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.587166556 podStartE2EDuration="2.587166556s" podCreationTimestamp="2025-10-03 15:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:30.580857892 +0000 UTC m=+5819.784201339" watchObservedRunningTime="2025-10-03 15:07:30.587166556 +0000 UTC m=+5819.790509983" Oct 03 15:07:36 crc kubenswrapper[4959]: I1003 15:07:36.685887 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:07:36 crc kubenswrapper[4959]: E1003 15:07:36.686921 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:07:38 crc kubenswrapper[4959]: I1003 15:07:38.978233 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.487791 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qxjgp"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.489633 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.492692 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.497692 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.501078 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxjgp"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.581209 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnrw6\" (UniqueName: \"kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.581265 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.581337 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.581542 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.629179 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.631007 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.634320 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.665278 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683171 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683278 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7rx9\" (UniqueName: \"kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683351 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683456 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnrw6\" (UniqueName: \"kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.683496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.693162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.695830 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.711917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.729763 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnrw6\" (UniqueName: \"kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6\") pod \"nova-cell0-cell-mapping-qxjgp\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.756809 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.761629 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.769453 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.785700 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.785824 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.785828 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.835068 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.838869 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.839052 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7rx9\" (UniqueName: \"kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.839202 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.839388 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.848329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.849845 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.849989 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.855492 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.877544 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.887914 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7rx9\" (UniqueName: \"kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9\") pod \"nova-api-0\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.932017 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.935613 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.938291 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.953200 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.953754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.954000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh29r\" (UniqueName: \"kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.954513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq4j6\" (UniqueName: \"kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.954699 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.955510 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmb87\" (UniqueName: \"kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.956019 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.956317 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.956894 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.957102 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.960369 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.963093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:39 crc kubenswrapper[4959]: I1003 15:07:39.982419 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.000470 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.000654 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060366 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060411 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060432 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060455 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060490 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh29r\" (UniqueName: \"kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq4j6\" (UniqueName: \"kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060562 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060588 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fphk\" (UniqueName: \"kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060616 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmb87\" (UniqueName: \"kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060683 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.060717 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.072594 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.073783 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.074131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.085307 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.106390 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.106408 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh29r\" (UniqueName: \"kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r\") pod \"nova-scheduler-0\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.106692 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.106900 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.114519 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq4j6\" (UniqueName: \"kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6\") pod \"nova-metadata-0\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.117447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmb87\" (UniqueName: \"kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87\") pod \"nova-cell1-novncproxy-0\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.162350 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.162462 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.162494 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fphk\" (UniqueName: \"kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.162540 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.162564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.163567 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.163642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.164075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.164678 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.187727 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fphk\" (UniqueName: \"kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk\") pod \"dnsmasq-dns-696f9966c7-bfs89\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.293833 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.306740 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.319552 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.328004 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.465389 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxjgp"] Oct 03 15:07:40 crc kubenswrapper[4959]: W1003 15:07:40.516161 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6d97872_2c24_4ee7_ae05_e599ef218ded.slice/crio-ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f WatchSource:0}: Error finding container ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f: Status 404 returned error can't find the container with id ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.559089 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:40 crc kubenswrapper[4959]: W1003 15:07:40.580653 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod627babb8_a51c_4fc3_8a41_6ae6877ebd73.slice/crio-31bdb83d5ccddd37a774a229ff87023a0e91858fceceeba957e9fcfb4a1bc097 WatchSource:0}: Error finding container 31bdb83d5ccddd37a774a229ff87023a0e91858fceceeba957e9fcfb4a1bc097: Status 404 returned error can't find the container with id 31bdb83d5ccddd37a774a229ff87023a0e91858fceceeba957e9fcfb4a1bc097 Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.695225 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerStarted","Data":"31bdb83d5ccddd37a774a229ff87023a0e91858fceceeba957e9fcfb4a1bc097"} Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.709704 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxjgp" event={"ID":"e6d97872-2c24-4ee7-ae05-e599ef218ded","Type":"ContainerStarted","Data":"ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f"} Oct 03 15:07:40 crc kubenswrapper[4959]: I1003 15:07:40.857063 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:07:40 crc kubenswrapper[4959]: W1003 15:07:40.881760 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dcab959_eef4_4d50_9d7e_1ae36849c889.slice/crio-3d0187dafa2551ab860390b927b518e3434c66fe04e99e921c7afdf0fda356de WatchSource:0}: Error finding container 3d0187dafa2551ab860390b927b518e3434c66fe04e99e921c7afdf0fda356de: Status 404 returned error can't find the container with id 3d0187dafa2551ab860390b927b518e3434c66fe04e99e921c7afdf0fda356de Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.023971 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gmgst"] Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.026464 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.029740 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.030256 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.036741 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:41 crc kubenswrapper[4959]: W1003 15:07:41.049475 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod631240ba_5382_4353_bb51_595569da2fb8.slice/crio-3be44e43426f155c914aed28a1df78f0247dc7fe4986c21c86bbaad8a97fcebb WatchSource:0}: Error finding container 3be44e43426f155c914aed28a1df78f0247dc7fe4986c21c86bbaad8a97fcebb: Status 404 returned error can't find the container with id 3be44e43426f155c914aed28a1df78f0247dc7fe4986c21c86bbaad8a97fcebb Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.069493 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gmgst"] Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.090613 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.090661 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.090775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.090824 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xwdh\" (UniqueName: \"kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.098211 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:07:41 crc kubenswrapper[4959]: W1003 15:07:41.106997 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3ece186_40f8_4d71_afe1_d2fc1653bd5a.slice/crio-24e3c74c226883cfa3408797d7398517c4106f9675c6de252eba2c3899effefe WatchSource:0}: Error finding container 24e3c74c226883cfa3408797d7398517c4106f9675c6de252eba2c3899effefe: Status 404 returned error can't find the container with id 24e3c74c226883cfa3408797d7398517c4106f9675c6de252eba2c3899effefe Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.118119 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.192869 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.192931 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xwdh\" (UniqueName: \"kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.192988 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.193014 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.206851 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.209338 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.213856 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.213924 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xwdh\" (UniqueName: \"kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh\") pod \"nova-cell1-conductor-db-sync-gmgst\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.364385 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.731341 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerStarted","Data":"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.731683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerStarted","Data":"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.731699 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerStarted","Data":"f08cc06865849e8c20962634b6d30c9b164aef3e342c80dc93fabc4b1e2caeb9"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.741668 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4dcab959-eef4-4d50-9d7e-1ae36849c889","Type":"ContainerStarted","Data":"011b44e5eb3ea9ec941ba1fa482024d8835f9b68364c5eddd0ba0d656936540f"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.741714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4dcab959-eef4-4d50-9d7e-1ae36849c889","Type":"ContainerStarted","Data":"3d0187dafa2551ab860390b927b518e3434c66fe04e99e921c7afdf0fda356de"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.747574 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"631240ba-5382-4353-bb51-595569da2fb8","Type":"ContainerStarted","Data":"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.747613 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"631240ba-5382-4353-bb51-595569da2fb8","Type":"ContainerStarted","Data":"3be44e43426f155c914aed28a1df78f0247dc7fe4986c21c86bbaad8a97fcebb"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.757557 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerStarted","Data":"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.757616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerStarted","Data":"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.759502 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxjgp" event={"ID":"e6d97872-2c24-4ee7-ae05-e599ef218ded","Type":"ContainerStarted","Data":"f86399ce1ead24b1b0f6d2633b1e9bc4f257535d260829ff27612e16f1254b07"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.767022 4959 generic.go:334] "Generic (PLEG): container finished" podID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerID="fa4e295f9888994678dfef497d1bcf1066db6e8a766b755ba7ad12f30e386db9" exitCode=0 Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.767072 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" event={"ID":"d3ece186-40f8-4d71-afe1-d2fc1653bd5a","Type":"ContainerDied","Data":"fa4e295f9888994678dfef497d1bcf1066db6e8a766b755ba7ad12f30e386db9"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.767098 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" event={"ID":"d3ece186-40f8-4d71-afe1-d2fc1653bd5a","Type":"ContainerStarted","Data":"24e3c74c226883cfa3408797d7398517c4106f9675c6de252eba2c3899effefe"} Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.884162 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gmgst"] Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.937499 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.937459617 podStartE2EDuration="2.937459617s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:41.884162473 +0000 UTC m=+5831.087505900" watchObservedRunningTime="2025-10-03 15:07:41.937459617 +0000 UTC m=+5831.140803034" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.961531 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.961498395 podStartE2EDuration="2.961498395s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:41.897082989 +0000 UTC m=+5831.100426396" watchObservedRunningTime="2025-10-03 15:07:41.961498395 +0000 UTC m=+5831.164841812" Oct 03 15:07:41 crc kubenswrapper[4959]: I1003 15:07:41.996606 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qxjgp" podStartSLOduration=2.996541593 podStartE2EDuration="2.996541593s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:41.970628199 +0000 UTC m=+5831.173971616" watchObservedRunningTime="2025-10-03 15:07:41.996541593 +0000 UTC m=+5831.199885010" Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.002613 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.00255881 podStartE2EDuration="3.00255881s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:41.993042948 +0000 UTC m=+5831.196386355" watchObservedRunningTime="2025-10-03 15:07:42.00255881 +0000 UTC m=+5831.205902227" Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.023645 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.023616665 podStartE2EDuration="3.023616665s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:42.009576492 +0000 UTC m=+5831.212919909" watchObservedRunningTime="2025-10-03 15:07:42.023616665 +0000 UTC m=+5831.226960082" Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.780931 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" event={"ID":"d3ece186-40f8-4d71-afe1-d2fc1653bd5a","Type":"ContainerStarted","Data":"2c5ca118997f0d8311b90f962b16584b4c0328746d303bb7a0792d4dd3954ed1"} Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.781512 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.784384 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gmgst" event={"ID":"f6224964-94b7-443c-abd6-4c101e4fa9fc","Type":"ContainerStarted","Data":"4e02e273de6ad29814726865a5dc6c0f6a073c121670d02c2f8ef57325111b01"} Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.784593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gmgst" event={"ID":"f6224964-94b7-443c-abd6-4c101e4fa9fc","Type":"ContainerStarted","Data":"a594626077ce232b4a57afe8a9279d8244c4c451f216153a7d1668c7adeab11a"} Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.812247 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" podStartSLOduration=3.812196062 podStartE2EDuration="3.812196062s" podCreationTimestamp="2025-10-03 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:42.803917419 +0000 UTC m=+5832.007260836" watchObservedRunningTime="2025-10-03 15:07:42.812196062 +0000 UTC m=+5832.015539469" Oct 03 15:07:42 crc kubenswrapper[4959]: I1003 15:07:42.828795 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gmgst" podStartSLOduration=2.828766877 podStartE2EDuration="2.828766877s" podCreationTimestamp="2025-10-03 15:07:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:42.826841819 +0000 UTC m=+5832.030185256" watchObservedRunningTime="2025-10-03 15:07:42.828766877 +0000 UTC m=+5832.032110294" Oct 03 15:07:44 crc kubenswrapper[4959]: I1003 15:07:44.823148 4959 generic.go:334] "Generic (PLEG): container finished" podID="f6224964-94b7-443c-abd6-4c101e4fa9fc" containerID="4e02e273de6ad29814726865a5dc6c0f6a073c121670d02c2f8ef57325111b01" exitCode=0 Oct 03 15:07:44 crc kubenswrapper[4959]: I1003 15:07:44.824251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gmgst" event={"ID":"f6224964-94b7-443c-abd6-4c101e4fa9fc","Type":"ContainerDied","Data":"4e02e273de6ad29814726865a5dc6c0f6a073c121670d02c2f8ef57325111b01"} Oct 03 15:07:45 crc kubenswrapper[4959]: I1003 15:07:45.294743 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:45 crc kubenswrapper[4959]: I1003 15:07:45.307744 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 15:07:45 crc kubenswrapper[4959]: I1003 15:07:45.320077 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:07:45 crc kubenswrapper[4959]: I1003 15:07:45.320942 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.219096 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.248366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xwdh\" (UniqueName: \"kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh\") pod \"f6224964-94b7-443c-abd6-4c101e4fa9fc\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.248471 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle\") pod \"f6224964-94b7-443c-abd6-4c101e4fa9fc\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.248579 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts\") pod \"f6224964-94b7-443c-abd6-4c101e4fa9fc\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.248738 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data\") pod \"f6224964-94b7-443c-abd6-4c101e4fa9fc\" (UID: \"f6224964-94b7-443c-abd6-4c101e4fa9fc\") " Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.255467 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh" (OuterVolumeSpecName: "kube-api-access-9xwdh") pod "f6224964-94b7-443c-abd6-4c101e4fa9fc" (UID: "f6224964-94b7-443c-abd6-4c101e4fa9fc"). InnerVolumeSpecName "kube-api-access-9xwdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.257325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts" (OuterVolumeSpecName: "scripts") pod "f6224964-94b7-443c-abd6-4c101e4fa9fc" (UID: "f6224964-94b7-443c-abd6-4c101e4fa9fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.275233 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6224964-94b7-443c-abd6-4c101e4fa9fc" (UID: "f6224964-94b7-443c-abd6-4c101e4fa9fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.276731 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data" (OuterVolumeSpecName: "config-data") pod "f6224964-94b7-443c-abd6-4c101e4fa9fc" (UID: "f6224964-94b7-443c-abd6-4c101e4fa9fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.352844 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.352895 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xwdh\" (UniqueName: \"kubernetes.io/projected/f6224964-94b7-443c-abd6-4c101e4fa9fc-kube-api-access-9xwdh\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.352912 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.352929 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6224964-94b7-443c-abd6-4c101e4fa9fc-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.848114 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6d97872-2c24-4ee7-ae05-e599ef218ded" containerID="f86399ce1ead24b1b0f6d2633b1e9bc4f257535d260829ff27612e16f1254b07" exitCode=0 Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.848242 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxjgp" event={"ID":"e6d97872-2c24-4ee7-ae05-e599ef218ded","Type":"ContainerDied","Data":"f86399ce1ead24b1b0f6d2633b1e9bc4f257535d260829ff27612e16f1254b07"} Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.851808 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gmgst" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.851920 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gmgst" event={"ID":"f6224964-94b7-443c-abd6-4c101e4fa9fc","Type":"ContainerDied","Data":"a594626077ce232b4a57afe8a9279d8244c4c451f216153a7d1668c7adeab11a"} Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.851949 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a594626077ce232b4a57afe8a9279d8244c4c451f216153a7d1668c7adeab11a" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.975091 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:07:46 crc kubenswrapper[4959]: E1003 15:07:46.975629 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6224964-94b7-443c-abd6-4c101e4fa9fc" containerName="nova-cell1-conductor-db-sync" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.975650 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6224964-94b7-443c-abd6-4c101e4fa9fc" containerName="nova-cell1-conductor-db-sync" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.975833 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6224964-94b7-443c-abd6-4c101e4fa9fc" containerName="nova-cell1-conductor-db-sync" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.976467 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.978548 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 15:07:46 crc kubenswrapper[4959]: I1003 15:07:46.987930 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.068836 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpbhp\" (UniqueName: \"kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.068961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.069051 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.170547 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpbhp\" (UniqueName: \"kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.170605 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.170706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.176847 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.177565 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.192462 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpbhp\" (UniqueName: \"kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp\") pod \"nova-cell1-conductor-0\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.303824 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:47 crc kubenswrapper[4959]: W1003 15:07:47.789755 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55d84893_19fb_4b35_a5fa_8e45af23ec3a.slice/crio-f98fd033b52d882adf827008ebe435d02a74778f4a192b0a39369f4f165ecc88 WatchSource:0}: Error finding container f98fd033b52d882adf827008ebe435d02a74778f4a192b0a39369f4f165ecc88: Status 404 returned error can't find the container with id f98fd033b52d882adf827008ebe435d02a74778f4a192b0a39369f4f165ecc88 Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.799382 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:07:47 crc kubenswrapper[4959]: I1003 15:07:47.878494 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"55d84893-19fb-4b35-a5fa-8e45af23ec3a","Type":"ContainerStarted","Data":"f98fd033b52d882adf827008ebe435d02a74778f4a192b0a39369f4f165ecc88"} Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.141641 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.194043 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle\") pod \"e6d97872-2c24-4ee7-ae05-e599ef218ded\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.194096 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data\") pod \"e6d97872-2c24-4ee7-ae05-e599ef218ded\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.194167 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnrw6\" (UniqueName: \"kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6\") pod \"e6d97872-2c24-4ee7-ae05-e599ef218ded\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.194281 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts\") pod \"e6d97872-2c24-4ee7-ae05-e599ef218ded\" (UID: \"e6d97872-2c24-4ee7-ae05-e599ef218ded\") " Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.200716 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6" (OuterVolumeSpecName: "kube-api-access-dnrw6") pod "e6d97872-2c24-4ee7-ae05-e599ef218ded" (UID: "e6d97872-2c24-4ee7-ae05-e599ef218ded"). InnerVolumeSpecName "kube-api-access-dnrw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.203371 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts" (OuterVolumeSpecName: "scripts") pod "e6d97872-2c24-4ee7-ae05-e599ef218ded" (UID: "e6d97872-2c24-4ee7-ae05-e599ef218ded"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.224852 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6d97872-2c24-4ee7-ae05-e599ef218ded" (UID: "e6d97872-2c24-4ee7-ae05-e599ef218ded"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.267356 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data" (OuterVolumeSpecName: "config-data") pod "e6d97872-2c24-4ee7-ae05-e599ef218ded" (UID: "e6d97872-2c24-4ee7-ae05-e599ef218ded"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.296714 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.296747 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.296778 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnrw6\" (UniqueName: \"kubernetes.io/projected/e6d97872-2c24-4ee7-ae05-e599ef218ded-kube-api-access-dnrw6\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.296789 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6d97872-2c24-4ee7-ae05-e599ef218ded-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.889427 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"55d84893-19fb-4b35-a5fa-8e45af23ec3a","Type":"ContainerStarted","Data":"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203"} Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.889835 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.891688 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxjgp" event={"ID":"e6d97872-2c24-4ee7-ae05-e599ef218ded","Type":"ContainerDied","Data":"ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f"} Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.891717 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecce7a9a2e3473f16e6a8640f7663e6efacb13299389d299613dade5310d1c3f" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.891760 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxjgp" Oct 03 15:07:48 crc kubenswrapper[4959]: I1003 15:07:48.927714 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.927695981 podStartE2EDuration="2.927695981s" podCreationTimestamp="2025-10-03 15:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:48.920491665 +0000 UTC m=+5838.123835092" watchObservedRunningTime="2025-10-03 15:07:48.927695981 +0000 UTC m=+5838.131039398" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.084969 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.085379 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="631240ba-5382-4353-bb51-595569da2fb8" containerName="nova-scheduler-scheduler" containerID="cri-o://e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286" gracePeriod=30 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.097966 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.098253 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-log" containerID="cri-o://1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" gracePeriod=30 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.098648 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-api" containerID="cri-o://cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" gracePeriod=30 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.110628 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.110886 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-log" containerID="cri-o://c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" gracePeriod=30 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.111014 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-metadata" containerID="cri-o://fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" gracePeriod=30 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.691830 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.706993 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.730234 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs\") pod \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.730338 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle\") pod \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.730415 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq4j6\" (UniqueName: \"kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6\") pod \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.730506 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data\") pod \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\" (UID: \"56685ee2-2c8a-4c92-ab43-5d18ac34250a\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.739060 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs" (OuterVolumeSpecName: "logs") pod "56685ee2-2c8a-4c92-ab43-5d18ac34250a" (UID: "56685ee2-2c8a-4c92-ab43-5d18ac34250a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.745604 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6" (OuterVolumeSpecName: "kube-api-access-kq4j6") pod "56685ee2-2c8a-4c92-ab43-5d18ac34250a" (UID: "56685ee2-2c8a-4c92-ab43-5d18ac34250a"). InnerVolumeSpecName "kube-api-access-kq4j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.762325 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56685ee2-2c8a-4c92-ab43-5d18ac34250a" (UID: "56685ee2-2c8a-4c92-ab43-5d18ac34250a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.765412 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data" (OuterVolumeSpecName: "config-data") pod "56685ee2-2c8a-4c92-ab43-5d18ac34250a" (UID: "56685ee2-2c8a-4c92-ab43-5d18ac34250a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.837274 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7rx9\" (UniqueName: \"kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9\") pod \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.837354 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle\") pod \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.837531 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data\") pod \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.837575 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs\") pod \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\" (UID: \"627babb8-a51c-4fc3-8a41-6ae6877ebd73\") " Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.837963 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs" (OuterVolumeSpecName: "logs") pod "627babb8-a51c-4fc3-8a41-6ae6877ebd73" (UID: "627babb8-a51c-4fc3-8a41-6ae6877ebd73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.838023 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.838152 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56685ee2-2c8a-4c92-ab43-5d18ac34250a-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.838248 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56685ee2-2c8a-4c92-ab43-5d18ac34250a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.838324 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq4j6\" (UniqueName: \"kubernetes.io/projected/56685ee2-2c8a-4c92-ab43-5d18ac34250a-kube-api-access-kq4j6\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.839769 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9" (OuterVolumeSpecName: "kube-api-access-d7rx9") pod "627babb8-a51c-4fc3-8a41-6ae6877ebd73" (UID: "627babb8-a51c-4fc3-8a41-6ae6877ebd73"). InnerVolumeSpecName "kube-api-access-d7rx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.860927 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data" (OuterVolumeSpecName: "config-data") pod "627babb8-a51c-4fc3-8a41-6ae6877ebd73" (UID: "627babb8-a51c-4fc3-8a41-6ae6877ebd73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.862244 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "627babb8-a51c-4fc3-8a41-6ae6877ebd73" (UID: "627babb8-a51c-4fc3-8a41-6ae6877ebd73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901045 4959 generic.go:334] "Generic (PLEG): container finished" podID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerID="fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" exitCode=0 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901098 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerDied","Data":"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901137 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerDied","Data":"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901154 4959 scope.go:117] "RemoveContainer" containerID="fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901151 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901106 4959 generic.go:334] "Generic (PLEG): container finished" podID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerID="c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" exitCode=143 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.901255 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"56685ee2-2c8a-4c92-ab43-5d18ac34250a","Type":"ContainerDied","Data":"f08cc06865849e8c20962634b6d30c9b164aef3e342c80dc93fabc4b1e2caeb9"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904639 4959 generic.go:334] "Generic (PLEG): container finished" podID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerID="cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" exitCode=0 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904701 4959 generic.go:334] "Generic (PLEG): container finished" podID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerID="1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" exitCode=143 Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904728 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerDied","Data":"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904908 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerDied","Data":"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.904937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"627babb8-a51c-4fc3-8a41-6ae6877ebd73","Type":"ContainerDied","Data":"31bdb83d5ccddd37a774a229ff87023a0e91858fceceeba957e9fcfb4a1bc097"} Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.927541 4959 scope.go:117] "RemoveContainer" containerID="c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.932807 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.945202 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.945254 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/627babb8-a51c-4fc3-8a41-6ae6877ebd73-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.945263 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7rx9\" (UniqueName: \"kubernetes.io/projected/627babb8-a51c-4fc3-8a41-6ae6877ebd73-kube-api-access-d7rx9\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.945274 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/627babb8-a51c-4fc3-8a41-6ae6877ebd73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.948502 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.969344 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.977457 4959 scope.go:117] "RemoveContainer" containerID="fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" Oct 03 15:07:49 crc kubenswrapper[4959]: E1003 15:07:49.981378 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd\": container with ID starting with fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd not found: ID does not exist" containerID="fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.981437 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd"} err="failed to get container status \"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd\": rpc error: code = NotFound desc = could not find container \"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd\": container with ID starting with fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd not found: ID does not exist" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.981471 4959 scope.go:117] "RemoveContainer" containerID="c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.983588 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:49 crc kubenswrapper[4959]: E1003 15:07:49.986341 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112\": container with ID starting with c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112 not found: ID does not exist" containerID="c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.986384 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112"} err="failed to get container status \"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112\": rpc error: code = NotFound desc = could not find container \"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112\": container with ID starting with c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112 not found: ID does not exist" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.986413 4959 scope.go:117] "RemoveContainer" containerID="fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.986854 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd"} err="failed to get container status \"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd\": rpc error: code = NotFound desc = could not find container \"fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd\": container with ID starting with fe45c7d45ee9e9e87c9f4bb9061d0bbf41a37453cc2986abadaf2997476af0fd not found: ID does not exist" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.986893 4959 scope.go:117] "RemoveContainer" containerID="c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.989195 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112"} err="failed to get container status \"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112\": rpc error: code = NotFound desc = could not find container \"c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112\": container with ID starting with c5701a30a7101a814551be9e434b90bd21a75179c0376afd2d8aae1d5ba55112 not found: ID does not exist" Oct 03 15:07:49 crc kubenswrapper[4959]: I1003 15:07:49.989244 4959 scope.go:117] "RemoveContainer" containerID="cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.000413 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.000921 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-log" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.000946 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-log" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.000971 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-metadata" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.000980 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-metadata" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.000997 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-log" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001008 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-log" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.001053 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-api" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001065 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-api" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.001080 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d97872-2c24-4ee7-ae05-e599ef218ded" containerName="nova-manage" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001090 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d97872-2c24-4ee7-ae05-e599ef218ded" containerName="nova-manage" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001378 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d97872-2c24-4ee7-ae05-e599ef218ded" containerName="nova-manage" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001410 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-log" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001442 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" containerName="nova-metadata-metadata" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001457 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-log" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.001477 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" containerName="nova-api-api" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.004055 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.009748 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.011505 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.012684 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.016458 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.017483 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.018651 4959 scope.go:117] "RemoveContainer" containerID="1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.024826 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.045164 4959 scope.go:117] "RemoveContainer" containerID="cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.045658 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2\": container with ID starting with cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2 not found: ID does not exist" containerID="cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.045703 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2"} err="failed to get container status \"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2\": rpc error: code = NotFound desc = could not find container \"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2\": container with ID starting with cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2 not found: ID does not exist" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.045755 4959 scope.go:117] "RemoveContainer" containerID="1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" Oct 03 15:07:50 crc kubenswrapper[4959]: E1003 15:07:50.046494 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5\": container with ID starting with 1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5 not found: ID does not exist" containerID="1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.046546 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5"} err="failed to get container status \"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5\": rpc error: code = NotFound desc = could not find container \"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5\": container with ID starting with 1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5 not found: ID does not exist" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.046603 4959 scope.go:117] "RemoveContainer" containerID="cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049062 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049159 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049218 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz2jt\" (UniqueName: \"kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049336 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049475 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049610 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049639 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2"} err="failed to get container status \"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2\": rpc error: code = NotFound desc = could not find container \"cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2\": container with ID starting with cdaf26fa85b781d61e8e599fcd1279d6de537f715bfda8da14691beaaffceda2 not found: ID does not exist" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.049674 4959 scope.go:117] "RemoveContainer" containerID="1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.050077 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.050153 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wqh9\" (UniqueName: \"kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.050664 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5"} err="failed to get container status \"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5\": rpc error: code = NotFound desc = could not find container \"1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5\": container with ID starting with 1ce9cac054ca51fc212b575d93f45992a4ae860e7a7acd2ee3e668078baa4fc5 not found: ID does not exist" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.151854 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.151929 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.151957 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.151983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wqh9\" (UniqueName: \"kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.152037 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.152057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.152079 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz2jt\" (UniqueName: \"kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.152096 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.152587 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.154317 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.155668 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.156226 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.156864 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.157258 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.175062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz2jt\" (UniqueName: \"kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt\") pod \"nova-api-0\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.180247 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wqh9\" (UniqueName: \"kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9\") pod \"nova-metadata-0\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.294805 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.304620 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.332584 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.334954 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.347420 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.457611 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.457907 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="dnsmasq-dns" containerID="cri-o://8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8" gracePeriod=10 Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.883935 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.931732 4959 generic.go:334] "Generic (PLEG): container finished" podID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerID="8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8" exitCode=0 Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.931805 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.931861 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" event={"ID":"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c","Type":"ContainerDied","Data":"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8"} Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.931893 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74df65d56c-6ch7g" event={"ID":"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c","Type":"ContainerDied","Data":"865ea30b7bb3afdca7a74fe26a3a403c8dfb839490767f741d8f60787bc222c2"} Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.931913 4959 scope.go:117] "RemoveContainer" containerID="8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.939048 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.943762 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.956393 4959 scope.go:117] "RemoveContainer" containerID="ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93" Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.979504 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb\") pod \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.979607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc\") pod \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.979649 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzv9n\" (UniqueName: \"kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n\") pod \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.980736 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config\") pod \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.980848 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb\") pod \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\" (UID: \"d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c\") " Oct 03 15:07:50 crc kubenswrapper[4959]: I1003 15:07:50.987465 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n" (OuterVolumeSpecName: "kube-api-access-pzv9n") pod "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" (UID: "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c"). InnerVolumeSpecName "kube-api-access-pzv9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.039861 4959 scope.go:117] "RemoveContainer" containerID="8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8" Oct 03 15:07:51 crc kubenswrapper[4959]: E1003 15:07:51.040292 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8\": container with ID starting with 8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8 not found: ID does not exist" containerID="8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.040423 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8"} err="failed to get container status \"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8\": rpc error: code = NotFound desc = could not find container \"8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8\": container with ID starting with 8e79c7f36486cf2fb1f70aebca3210ef935add4cca400d1ed6dc681b7fbab6b8 not found: ID does not exist" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.040535 4959 scope.go:117] "RemoveContainer" containerID="ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93" Oct 03 15:07:51 crc kubenswrapper[4959]: E1003 15:07:51.042328 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93\": container with ID starting with ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93 not found: ID does not exist" containerID="ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.042362 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93"} err="failed to get container status \"ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93\": rpc error: code = NotFound desc = could not find container \"ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93\": container with ID starting with ef82fef3fefe391d0fbe6c72d2e5ac2f0a016484a84909e4cc79affed24eba93 not found: ID does not exist" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.053093 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" (UID: "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.059832 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config" (OuterVolumeSpecName: "config") pod "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" (UID: "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.062818 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" (UID: "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.066994 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" (UID: "d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.068467 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:07:51 crc kubenswrapper[4959]: W1003 15:07:51.072363 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08d4a4ea_d912_4afc_9df9_aca93bd06aa9.slice/crio-4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a WatchSource:0}: Error finding container 4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a: Status 404 returned error can't find the container with id 4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.083686 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.083717 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.083729 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.083738 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.083748 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzv9n\" (UniqueName: \"kubernetes.io/projected/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c-kube-api-access-pzv9n\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.380285 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.387624 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74df65d56c-6ch7g"] Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.694559 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:07:51 crc kubenswrapper[4959]: E1003 15:07:51.695055 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.697996 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56685ee2-2c8a-4c92-ab43-5d18ac34250a" path="/var/lib/kubelet/pods/56685ee2-2c8a-4c92-ab43-5d18ac34250a/volumes" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.698723 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="627babb8-a51c-4fc3-8a41-6ae6877ebd73" path="/var/lib/kubelet/pods/627babb8-a51c-4fc3-8a41-6ae6877ebd73/volumes" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.699478 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" path="/var/lib/kubelet/pods/d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c/volumes" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.942426 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerStarted","Data":"eed0267e026a834788e77b5abf7f511f02ca3f55bfa538c2ab88020c216a8a22"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.942476 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerStarted","Data":"5876ccf01e94c8c4a7f84a47ca061eadd20ab74383c3202dd5510873188c2ea6"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.942491 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerStarted","Data":"4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.944836 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerStarted","Data":"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.944871 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerStarted","Data":"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.944885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerStarted","Data":"872fdf4f3881ad2c262ed7fb280ade8aef10f58d69caaf77563637273d0ba55e"} Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.964888 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.964866619 podStartE2EDuration="2.964866619s" podCreationTimestamp="2025-10-03 15:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:51.958537353 +0000 UTC m=+5841.161880770" watchObservedRunningTime="2025-10-03 15:07:51.964866619 +0000 UTC m=+5841.168210036" Oct 03 15:07:51 crc kubenswrapper[4959]: I1003 15:07:51.982238 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.982221073 podStartE2EDuration="2.982221073s" podCreationTimestamp="2025-10-03 15:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:51.978082422 +0000 UTC m=+5841.181425829" watchObservedRunningTime="2025-10-03 15:07:51.982221073 +0000 UTC m=+5841.185564490" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.338910 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.860878 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.929673 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh29r\" (UniqueName: \"kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r\") pod \"631240ba-5382-4353-bb51-595569da2fb8\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.929811 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle\") pod \"631240ba-5382-4353-bb51-595569da2fb8\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.929947 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data\") pod \"631240ba-5382-4353-bb51-595569da2fb8\" (UID: \"631240ba-5382-4353-bb51-595569da2fb8\") " Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.947676 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kmnd8"] Oct 03 15:07:52 crc kubenswrapper[4959]: E1003 15:07:52.948637 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631240ba-5382-4353-bb51-595569da2fb8" containerName="nova-scheduler-scheduler" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.948670 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="631240ba-5382-4353-bb51-595569da2fb8" containerName="nova-scheduler-scheduler" Oct 03 15:07:52 crc kubenswrapper[4959]: E1003 15:07:52.948735 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="dnsmasq-dns" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.948745 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="dnsmasq-dns" Oct 03 15:07:52 crc kubenswrapper[4959]: E1003 15:07:52.948783 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="init" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.948791 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="init" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.949150 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="631240ba-5382-4353-bb51-595569da2fb8" containerName="nova-scheduler-scheduler" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.949193 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d8c4df-a879-4a83-9f3e-8b1a4b6d0f0c" containerName="dnsmasq-dns" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.950672 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.952132 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r" (OuterVolumeSpecName: "kube-api-access-rh29r") pod "631240ba-5382-4353-bb51-595569da2fb8" (UID: "631240ba-5382-4353-bb51-595569da2fb8"). InnerVolumeSpecName "kube-api-access-rh29r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.953395 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.953662 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.975889 4959 generic.go:334] "Generic (PLEG): container finished" podID="631240ba-5382-4353-bb51-595569da2fb8" containerID="e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286" exitCode=0 Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.976366 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"631240ba-5382-4353-bb51-595569da2fb8","Type":"ContainerDied","Data":"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286"} Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.976446 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"631240ba-5382-4353-bb51-595569da2fb8","Type":"ContainerDied","Data":"3be44e43426f155c914aed28a1df78f0247dc7fe4986c21c86bbaad8a97fcebb"} Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.976471 4959 scope.go:117] "RemoveContainer" containerID="e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.976847 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:52 crc kubenswrapper[4959]: I1003 15:07:52.978380 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data" (OuterVolumeSpecName: "config-data") pod "631240ba-5382-4353-bb51-595569da2fb8" (UID: "631240ba-5382-4353-bb51-595569da2fb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.002300 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmnd8"] Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031229 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74nwl\" (UniqueName: \"kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031281 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031547 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031689 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031791 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.031811 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh29r\" (UniqueName: \"kubernetes.io/projected/631240ba-5382-4353-bb51-595569da2fb8-kube-api-access-rh29r\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.034886 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "631240ba-5382-4353-bb51-595569da2fb8" (UID: "631240ba-5382-4353-bb51-595569da2fb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.047615 4959 scope.go:117] "RemoveContainer" containerID="e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286" Oct 03 15:07:53 crc kubenswrapper[4959]: E1003 15:07:53.048719 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286\": container with ID starting with e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286 not found: ID does not exist" containerID="e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.048766 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286"} err="failed to get container status \"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286\": rpc error: code = NotFound desc = could not find container \"e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286\": container with ID starting with e3a8fef63003eead225eb94555f99730dfafc7a571ba3974283323f74cd3a286 not found: ID does not exist" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.132938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.132990 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.133073 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74nwl\" (UniqueName: \"kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.133095 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.133156 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631240ba-5382-4353-bb51-595569da2fb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.137181 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.137265 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.138047 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.150068 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74nwl\" (UniqueName: \"kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl\") pod \"nova-cell1-cell-mapping-kmnd8\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.309963 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.320485 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.331038 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.334371 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.335727 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.341846 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.352316 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.437384 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.437716 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.437784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hp66\" (UniqueName: \"kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.539882 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.540032 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.540134 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hp66\" (UniqueName: \"kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.547183 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.549620 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.565520 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hp66\" (UniqueName: \"kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66\") pod \"nova-scheduler-0\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.603256 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmnd8"] Oct 03 15:07:53 crc kubenswrapper[4959]: W1003 15:07:53.604767 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d0c23d_fd4c_4027_9e06_7067e42c0c27.slice/crio-8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa WatchSource:0}: Error finding container 8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa: Status 404 returned error can't find the container with id 8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.654767 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.700387 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631240ba-5382-4353-bb51-595569da2fb8" path="/var/lib/kubelet/pods/631240ba-5382-4353-bb51-595569da2fb8/volumes" Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.988959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmnd8" event={"ID":"b6d0c23d-fd4c-4027-9e06-7067e42c0c27","Type":"ContainerStarted","Data":"eb1b282761a352975a2938dfa259624fd47ce90a79367510d6fec7b4dcb43e10"} Oct 03 15:07:53 crc kubenswrapper[4959]: I1003 15:07:53.989385 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmnd8" event={"ID":"b6d0c23d-fd4c-4027-9e06-7067e42c0c27","Type":"ContainerStarted","Data":"8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa"} Oct 03 15:07:54 crc kubenswrapper[4959]: I1003 15:07:54.018182 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kmnd8" podStartSLOduration=2.01815537 podStartE2EDuration="2.01815537s" podCreationTimestamp="2025-10-03 15:07:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:54.005770627 +0000 UTC m=+5843.209114034" watchObservedRunningTime="2025-10-03 15:07:54.01815537 +0000 UTC m=+5843.221498817" Oct 03 15:07:54 crc kubenswrapper[4959]: I1003 15:07:54.114161 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:07:54 crc kubenswrapper[4959]: W1003 15:07:54.121084 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e852434_86c5_4d6a_be5c_3231e2266e04.slice/crio-0e4fe2395a78a7d2ec6298839c82ed25110dbb056b941db79c3196512f4488c7 WatchSource:0}: Error finding container 0e4fe2395a78a7d2ec6298839c82ed25110dbb056b941db79c3196512f4488c7: Status 404 returned error can't find the container with id 0e4fe2395a78a7d2ec6298839c82ed25110dbb056b941db79c3196512f4488c7 Oct 03 15:07:55 crc kubenswrapper[4959]: I1003 15:07:55.003429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e852434-86c5-4d6a-be5c-3231e2266e04","Type":"ContainerStarted","Data":"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395"} Oct 03 15:07:55 crc kubenswrapper[4959]: I1003 15:07:55.003723 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e852434-86c5-4d6a-be5c-3231e2266e04","Type":"ContainerStarted","Data":"0e4fe2395a78a7d2ec6298839c82ed25110dbb056b941db79c3196512f4488c7"} Oct 03 15:07:55 crc kubenswrapper[4959]: I1003 15:07:55.027715 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.027691423 podStartE2EDuration="2.027691423s" podCreationTimestamp="2025-10-03 15:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:07:55.018698493 +0000 UTC m=+5844.222041910" watchObservedRunningTime="2025-10-03 15:07:55.027691423 +0000 UTC m=+5844.231034870" Oct 03 15:07:55 crc kubenswrapper[4959]: I1003 15:07:55.335660 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:07:55 crc kubenswrapper[4959]: I1003 15:07:55.336059 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:07:58 crc kubenswrapper[4959]: I1003 15:07:58.655830 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 15:07:59 crc kubenswrapper[4959]: I1003 15:07:59.045815 4959 generic.go:334] "Generic (PLEG): container finished" podID="b6d0c23d-fd4c-4027-9e06-7067e42c0c27" containerID="eb1b282761a352975a2938dfa259624fd47ce90a79367510d6fec7b4dcb43e10" exitCode=0 Oct 03 15:07:59 crc kubenswrapper[4959]: I1003 15:07:59.045903 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmnd8" event={"ID":"b6d0c23d-fd4c-4027-9e06-7067e42c0c27","Type":"ContainerDied","Data":"eb1b282761a352975a2938dfa259624fd47ce90a79367510d6fec7b4dcb43e10"} Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.336471 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.336906 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.349146 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.350463 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.479415 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.573558 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data\") pod \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.573621 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle\") pod \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.573732 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts\") pod \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.573872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74nwl\" (UniqueName: \"kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl\") pod \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\" (UID: \"b6d0c23d-fd4c-4027-9e06-7067e42c0c27\") " Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.578414 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl" (OuterVolumeSpecName: "kube-api-access-74nwl") pod "b6d0c23d-fd4c-4027-9e06-7067e42c0c27" (UID: "b6d0c23d-fd4c-4027-9e06-7067e42c0c27"). InnerVolumeSpecName "kube-api-access-74nwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.578456 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts" (OuterVolumeSpecName: "scripts") pod "b6d0c23d-fd4c-4027-9e06-7067e42c0c27" (UID: "b6d0c23d-fd4c-4027-9e06-7067e42c0c27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.608346 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d0c23d-fd4c-4027-9e06-7067e42c0c27" (UID: "b6d0c23d-fd4c-4027-9e06-7067e42c0c27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.616374 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data" (OuterVolumeSpecName: "config-data") pod "b6d0c23d-fd4c-4027-9e06-7067e42c0c27" (UID: "b6d0c23d-fd4c-4027-9e06-7067e42c0c27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.675885 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74nwl\" (UniqueName: \"kubernetes.io/projected/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-kube-api-access-74nwl\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.676146 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.676285 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:00 crc kubenswrapper[4959]: I1003 15:08:00.676404 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d0c23d-fd4c-4027-9e06-7067e42c0c27-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.080857 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kmnd8" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.081022 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kmnd8" event={"ID":"b6d0c23d-fd4c-4027-9e06-7067e42c0c27","Type":"ContainerDied","Data":"8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa"} Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.081210 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a537bdab36fd72574ae190cee267ea5f278dadb3ac7efcd75ca41fc11f553fa" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.242034 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.248414 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.248629 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9e852434-86c5-4d6a-be5c-3231e2266e04" containerName="nova-scheduler-scheduler" containerID="cri-o://6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395" gracePeriod=30 Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.316956 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.317211 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-log" containerID="cri-o://5876ccf01e94c8c4a7f84a47ca061eadd20ab74383c3202dd5510873188c2ea6" gracePeriod=30 Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.317397 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-metadata" containerID="cri-o://eed0267e026a834788e77b5abf7f511f02ca3f55bfa538c2ab88020c216a8a22" gracePeriod=30 Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.321827 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": EOF" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.322578 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.71:8775/\": EOF" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.431394 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:01 crc kubenswrapper[4959]: I1003 15:08:01.431428 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:02 crc kubenswrapper[4959]: I1003 15:08:02.087555 4959 generic.go:334] "Generic (PLEG): container finished" podID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerID="5876ccf01e94c8c4a7f84a47ca061eadd20ab74383c3202dd5510873188c2ea6" exitCode=143 Oct 03 15:08:02 crc kubenswrapper[4959]: I1003 15:08:02.087611 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerDied","Data":"5876ccf01e94c8c4a7f84a47ca061eadd20ab74383c3202dd5510873188c2ea6"} Oct 03 15:08:02 crc kubenswrapper[4959]: I1003 15:08:02.087778 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-log" containerID="cri-o://696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133" gracePeriod=30 Oct 03 15:08:02 crc kubenswrapper[4959]: I1003 15:08:02.087894 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-api" containerID="cri-o://3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91" gracePeriod=30 Oct 03 15:08:02 crc kubenswrapper[4959]: I1003 15:08:02.685795 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:08:02 crc kubenswrapper[4959]: E1003 15:08:02.686066 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:08:03 crc kubenswrapper[4959]: I1003 15:08:03.104854 4959 generic.go:334] "Generic (PLEG): container finished" podID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerID="696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133" exitCode=143 Oct 03 15:08:03 crc kubenswrapper[4959]: I1003 15:08:03.104926 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerDied","Data":"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133"} Oct 03 15:08:05 crc kubenswrapper[4959]: I1003 15:08:05.929106 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.078010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle\") pod \"9e852434-86c5-4d6a-be5c-3231e2266e04\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.078095 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data\") pod \"9e852434-86c5-4d6a-be5c-3231e2266e04\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.078174 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hp66\" (UniqueName: \"kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66\") pod \"9e852434-86c5-4d6a-be5c-3231e2266e04\" (UID: \"9e852434-86c5-4d6a-be5c-3231e2266e04\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.086277 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66" (OuterVolumeSpecName: "kube-api-access-6hp66") pod "9e852434-86c5-4d6a-be5c-3231e2266e04" (UID: "9e852434-86c5-4d6a-be5c-3231e2266e04"). InnerVolumeSpecName "kube-api-access-6hp66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.109764 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e852434-86c5-4d6a-be5c-3231e2266e04" (UID: "9e852434-86c5-4d6a-be5c-3231e2266e04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.114859 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data" (OuterVolumeSpecName: "config-data") pod "9e852434-86c5-4d6a-be5c-3231e2266e04" (UID: "9e852434-86c5-4d6a-be5c-3231e2266e04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.135523 4959 generic.go:334] "Generic (PLEG): container finished" podID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerID="eed0267e026a834788e77b5abf7f511f02ca3f55bfa538c2ab88020c216a8a22" exitCode=0 Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.135570 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerDied","Data":"eed0267e026a834788e77b5abf7f511f02ca3f55bfa538c2ab88020c216a8a22"} Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.135595 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"08d4a4ea-d912-4afc-9df9-aca93bd06aa9","Type":"ContainerDied","Data":"4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a"} Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.135605 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4208ee0b7c502b755208f4cbdebfc4b9de0e00589170290157346b02c537283a" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.137257 4959 generic.go:334] "Generic (PLEG): container finished" podID="9e852434-86c5-4d6a-be5c-3231e2266e04" containerID="6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395" exitCode=0 Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.137280 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e852434-86c5-4d6a-be5c-3231e2266e04","Type":"ContainerDied","Data":"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395"} Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.137295 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e852434-86c5-4d6a-be5c-3231e2266e04","Type":"ContainerDied","Data":"0e4fe2395a78a7d2ec6298839c82ed25110dbb056b941db79c3196512f4488c7"} Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.137310 4959 scope.go:117] "RemoveContainer" containerID="6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.137416 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.165582 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.179794 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hp66\" (UniqueName: \"kubernetes.io/projected/9e852434-86c5-4d6a-be5c-3231e2266e04-kube-api-access-6hp66\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.179834 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.179849 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e852434-86c5-4d6a-be5c-3231e2266e04-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.198366 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.206542 4959 scope.go:117] "RemoveContainer" containerID="6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395" Oct 03 15:08:06 crc kubenswrapper[4959]: E1003 15:08:06.207087 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395\": container with ID starting with 6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395 not found: ID does not exist" containerID="6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.207136 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395"} err="failed to get container status \"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395\": rpc error: code = NotFound desc = could not find container \"6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395\": container with ID starting with 6d921ddcaf554cdfa2b8744618f1c911af541c0d55e68bf44731903ec1c16395 not found: ID does not exist" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.213466 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.225796 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:06 crc kubenswrapper[4959]: E1003 15:08:06.227145 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-metadata" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227172 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-metadata" Oct 03 15:08:06 crc kubenswrapper[4959]: E1003 15:08:06.227205 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e852434-86c5-4d6a-be5c-3231e2266e04" containerName="nova-scheduler-scheduler" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227213 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e852434-86c5-4d6a-be5c-3231e2266e04" containerName="nova-scheduler-scheduler" Oct 03 15:08:06 crc kubenswrapper[4959]: E1003 15:08:06.227224 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d0c23d-fd4c-4027-9e06-7067e42c0c27" containerName="nova-manage" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227233 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d0c23d-fd4c-4027-9e06-7067e42c0c27" containerName="nova-manage" Oct 03 15:08:06 crc kubenswrapper[4959]: E1003 15:08:06.227246 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-log" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227255 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-log" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227455 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-log" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227480 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d0c23d-fd4c-4027-9e06-7067e42c0c27" containerName="nova-manage" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227491 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" containerName="nova-metadata-metadata" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.227511 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e852434-86c5-4d6a-be5c-3231e2266e04" containerName="nova-scheduler-scheduler" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.228585 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.230328 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.244771 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.281229 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle\") pod \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.281438 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs\") pod \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.281517 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data\") pod \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.281811 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wqh9\" (UniqueName: \"kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9\") pod \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\" (UID: \"08d4a4ea-d912-4afc-9df9-aca93bd06aa9\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.281871 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs" (OuterVolumeSpecName: "logs") pod "08d4a4ea-d912-4afc-9df9-aca93bd06aa9" (UID: "08d4a4ea-d912-4afc-9df9-aca93bd06aa9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.282405 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62wj\" (UniqueName: \"kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.282701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.282833 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.282966 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.284285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9" (OuterVolumeSpecName: "kube-api-access-2wqh9") pod "08d4a4ea-d912-4afc-9df9-aca93bd06aa9" (UID: "08d4a4ea-d912-4afc-9df9-aca93bd06aa9"). InnerVolumeSpecName "kube-api-access-2wqh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.307410 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08d4a4ea-d912-4afc-9df9-aca93bd06aa9" (UID: "08d4a4ea-d912-4afc-9df9-aca93bd06aa9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.307435 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data" (OuterVolumeSpecName: "config-data") pod "08d4a4ea-d912-4afc-9df9-aca93bd06aa9" (UID: "08d4a4ea-d912-4afc-9df9-aca93bd06aa9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.384555 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62wj\" (UniqueName: \"kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.384901 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.385038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.385145 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.385249 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wqh9\" (UniqueName: \"kubernetes.io/projected/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-kube-api-access-2wqh9\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.385344 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08d4a4ea-d912-4afc-9df9-aca93bd06aa9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.389669 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.390246 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.406725 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62wj\" (UniqueName: \"kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj\") pod \"nova-scheduler-0\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.549937 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.844350 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.994767 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data\") pod \"a84625a4-8033-401d-b1f3-1eef6475e78c\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.994904 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz2jt\" (UniqueName: \"kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt\") pod \"a84625a4-8033-401d-b1f3-1eef6475e78c\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.994997 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs\") pod \"a84625a4-8033-401d-b1f3-1eef6475e78c\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.995052 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle\") pod \"a84625a4-8033-401d-b1f3-1eef6475e78c\" (UID: \"a84625a4-8033-401d-b1f3-1eef6475e78c\") " Oct 03 15:08:06 crc kubenswrapper[4959]: I1003 15:08:06.995939 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs" (OuterVolumeSpecName: "logs") pod "a84625a4-8033-401d-b1f3-1eef6475e78c" (UID: "a84625a4-8033-401d-b1f3-1eef6475e78c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.038223 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt" (OuterVolumeSpecName: "kube-api-access-gz2jt") pod "a84625a4-8033-401d-b1f3-1eef6475e78c" (UID: "a84625a4-8033-401d-b1f3-1eef6475e78c"). InnerVolumeSpecName "kube-api-access-gz2jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.043280 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data" (OuterVolumeSpecName: "config-data") pod "a84625a4-8033-401d-b1f3-1eef6475e78c" (UID: "a84625a4-8033-401d-b1f3-1eef6475e78c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.043854 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a84625a4-8033-401d-b1f3-1eef6475e78c" (UID: "a84625a4-8033-401d-b1f3-1eef6475e78c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.074284 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: W1003 15:08:07.077417 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd456263_94c7_4963_97a4_fbaa389c2a9c.slice/crio-3b8560c4c4e9c16e45af4c1784b62615951d05f460975f180fe738f82b29a8d1 WatchSource:0}: Error finding container 3b8560c4c4e9c16e45af4c1784b62615951d05f460975f180fe738f82b29a8d1: Status 404 returned error can't find the container with id 3b8560c4c4e9c16e45af4c1784b62615951d05f460975f180fe738f82b29a8d1 Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.097812 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.097855 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz2jt\" (UniqueName: \"kubernetes.io/projected/a84625a4-8033-401d-b1f3-1eef6475e78c-kube-api-access-gz2jt\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.097868 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84625a4-8033-401d-b1f3-1eef6475e78c-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.097881 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84625a4-8033-401d-b1f3-1eef6475e78c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.149262 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd456263-94c7-4963-97a4-fbaa389c2a9c","Type":"ContainerStarted","Data":"3b8560c4c4e9c16e45af4c1784b62615951d05f460975f180fe738f82b29a8d1"} Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.152539 4959 generic.go:334] "Generic (PLEG): container finished" podID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerID="3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91" exitCode=0 Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.152613 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerDied","Data":"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91"} Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.152632 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a84625a4-8033-401d-b1f3-1eef6475e78c","Type":"ContainerDied","Data":"872fdf4f3881ad2c262ed7fb280ade8aef10f58d69caaf77563637273d0ba55e"} Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.152630 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.152668 4959 scope.go:117] "RemoveContainer" containerID="3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.155618 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.197865 4959 scope.go:117] "RemoveContainer" containerID="696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.207611 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.224979 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.237555 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: E1003 15:08:07.238064 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-log" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.238087 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-log" Oct 03 15:08:07 crc kubenswrapper[4959]: E1003 15:08:07.238125 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-api" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.238134 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-api" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.238402 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-api" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.238431 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" containerName="nova-api-log" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.239660 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.244303 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.249487 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.261998 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.271511 4959 scope.go:117] "RemoveContainer" containerID="3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91" Oct 03 15:08:07 crc kubenswrapper[4959]: E1003 15:08:07.273352 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91\": container with ID starting with 3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91 not found: ID does not exist" containerID="3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.273390 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91"} err="failed to get container status \"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91\": rpc error: code = NotFound desc = could not find container \"3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91\": container with ID starting with 3951a02de4d2d50d1853dfe6d081847c98b349763f2e8bec252e1be3e90baa91 not found: ID does not exist" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.273413 4959 scope.go:117] "RemoveContainer" containerID="696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133" Oct 03 15:08:07 crc kubenswrapper[4959]: E1003 15:08:07.273758 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133\": container with ID starting with 696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133 not found: ID does not exist" containerID="696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.273786 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133"} err="failed to get container status \"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133\": rpc error: code = NotFound desc = could not find container \"696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133\": container with ID starting with 696c2f3bb1c61b32f918c9df677aa1a611e7cfe166e1218e1735aa35f023c133 not found: ID does not exist" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.285086 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.294358 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.296264 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.300930 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.311221 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405514 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405604 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs9xx\" (UniqueName: \"kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405780 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.405980 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.406024 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.406105 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndfbv\" (UniqueName: \"kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507067 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507109 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs9xx\" (UniqueName: \"kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507149 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507176 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507222 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507287 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.507330 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndfbv\" (UniqueName: \"kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.508373 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.508844 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.514092 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.520835 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.520916 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.522355 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.523654 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndfbv\" (UniqueName: \"kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv\") pod \"nova-metadata-0\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.524649 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs9xx\" (UniqueName: \"kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx\") pod \"nova-api-0\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.572559 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.621735 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.699155 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d4a4ea-d912-4afc-9df9-aca93bd06aa9" path="/var/lib/kubelet/pods/08d4a4ea-d912-4afc-9df9-aca93bd06aa9/volumes" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.700274 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e852434-86c5-4d6a-be5c-3231e2266e04" path="/var/lib/kubelet/pods/9e852434-86c5-4d6a-be5c-3231e2266e04/volumes" Oct 03 15:08:07 crc kubenswrapper[4959]: I1003 15:08:07.701163 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a84625a4-8033-401d-b1f3-1eef6475e78c" path="/var/lib/kubelet/pods/a84625a4-8033-401d-b1f3-1eef6475e78c/volumes" Oct 03 15:08:08 crc kubenswrapper[4959]: W1003 15:08:08.037791 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d31aaca_d388_4930_9efe_35718e9898f5.slice/crio-924f5ec3147ea8dfe588e6a6cfde8f7d7b534f3707aa62c8ed86cc6d018f8960 WatchSource:0}: Error finding container 924f5ec3147ea8dfe588e6a6cfde8f7d7b534f3707aa62c8ed86cc6d018f8960: Status 404 returned error can't find the container with id 924f5ec3147ea8dfe588e6a6cfde8f7d7b534f3707aa62c8ed86cc6d018f8960 Oct 03 15:08:08 crc kubenswrapper[4959]: I1003 15:08:08.040443 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:08:08 crc kubenswrapper[4959]: I1003 15:08:08.172836 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:08:08 crc kubenswrapper[4959]: I1003 15:08:08.190249 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd456263-94c7-4963-97a4-fbaa389c2a9c","Type":"ContainerStarted","Data":"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4"} Oct 03 15:08:08 crc kubenswrapper[4959]: I1003 15:08:08.193678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerStarted","Data":"924f5ec3147ea8dfe588e6a6cfde8f7d7b534f3707aa62c8ed86cc6d018f8960"} Oct 03 15:08:08 crc kubenswrapper[4959]: I1003 15:08:08.216632 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.216602523 podStartE2EDuration="2.216602523s" podCreationTimestamp="2025-10-03 15:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:08:08.213472197 +0000 UTC m=+5857.416815644" watchObservedRunningTime="2025-10-03 15:08:08.216602523 +0000 UTC m=+5857.419945980" Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.202727 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerStarted","Data":"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec"} Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.203008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerStarted","Data":"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955"} Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.203017 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerStarted","Data":"0e7d26d05aa4ec888d5d589ff9c87fccbacd0656094495d10f5c08b808709787"} Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.206927 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerStarted","Data":"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c"} Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.206971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerStarted","Data":"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0"} Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.224840 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.224819633 podStartE2EDuration="2.224819633s" podCreationTimestamp="2025-10-03 15:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:08:09.222336472 +0000 UTC m=+5858.425679899" watchObservedRunningTime="2025-10-03 15:08:09.224819633 +0000 UTC m=+5858.428163050" Oct 03 15:08:09 crc kubenswrapper[4959]: I1003 15:08:09.255249 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.255222987 podStartE2EDuration="2.255222987s" podCreationTimestamp="2025-10-03 15:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:08:09.237913523 +0000 UTC m=+5858.441256980" watchObservedRunningTime="2025-10-03 15:08:09.255222987 +0000 UTC m=+5858.458566414" Oct 03 15:08:11 crc kubenswrapper[4959]: I1003 15:08:11.551071 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 15:08:12 crc kubenswrapper[4959]: I1003 15:08:12.572638 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:08:12 crc kubenswrapper[4959]: I1003 15:08:12.572710 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:08:13 crc kubenswrapper[4959]: I1003 15:08:13.685971 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:08:13 crc kubenswrapper[4959]: E1003 15:08:13.686473 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:08:16 crc kubenswrapper[4959]: I1003 15:08:16.550856 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 15:08:16 crc kubenswrapper[4959]: I1003 15:08:16.579298 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 15:08:17 crc kubenswrapper[4959]: I1003 15:08:17.327044 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 15:08:17 crc kubenswrapper[4959]: I1003 15:08:17.572875 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:08:17 crc kubenswrapper[4959]: I1003 15:08:17.572975 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:08:17 crc kubenswrapper[4959]: I1003 15:08:17.623425 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:08:17 crc kubenswrapper[4959]: I1003 15:08:17.623479 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:08:18 crc kubenswrapper[4959]: I1003 15:08:18.657423 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:18 crc kubenswrapper[4959]: I1003 15:08:18.657476 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:18 crc kubenswrapper[4959]: I1003 15:08:18.699498 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:18 crc kubenswrapper[4959]: I1003 15:08:18.699498 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.577018 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.577760 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.581360 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.582119 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.632007 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.632887 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.647592 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.650618 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 15:08:27 crc kubenswrapper[4959]: I1003 15:08:27.704376 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:08:27 crc kubenswrapper[4959]: E1003 15:08:27.704722 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.417679 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.421061 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.655277 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.656902 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.683648 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.770643 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.771154 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbt5k\" (UniqueName: \"kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.771410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.771450 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.771643 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.873252 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbt5k\" (UniqueName: \"kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.873314 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.873354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.873382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.873444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.874208 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.874271 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.874387 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.874463 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.895959 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbt5k\" (UniqueName: \"kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k\") pod \"dnsmasq-dns-64b8d7d4fc-ggxmr\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:28 crc kubenswrapper[4959]: I1003 15:08:28.979349 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:29 crc kubenswrapper[4959]: I1003 15:08:29.482149 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:08:30 crc kubenswrapper[4959]: I1003 15:08:30.442333 4959 generic.go:334] "Generic (PLEG): container finished" podID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerID="ef0e6ba0e15f69530d4ca8bde7b3b2664560364a20f32f7a16de743c888d6169" exitCode=0 Oct 03 15:08:30 crc kubenswrapper[4959]: I1003 15:08:30.442412 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" event={"ID":"6c6c71fd-235e-47d6-8220-bfbd1683d915","Type":"ContainerDied","Data":"ef0e6ba0e15f69530d4ca8bde7b3b2664560364a20f32f7a16de743c888d6169"} Oct 03 15:08:30 crc kubenswrapper[4959]: I1003 15:08:30.442642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" event={"ID":"6c6c71fd-235e-47d6-8220-bfbd1683d915","Type":"ContainerStarted","Data":"a8c87eae8e69d30268ccd218465ab1f739e41e79a259eb3ef7b8eaf446b578db"} Oct 03 15:08:31 crc kubenswrapper[4959]: I1003 15:08:31.451596 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" event={"ID":"6c6c71fd-235e-47d6-8220-bfbd1683d915","Type":"ContainerStarted","Data":"5a08731f8653d814f3ddd228af571530adba31c6da26e58ae4683fab0451bc7e"} Oct 03 15:08:31 crc kubenswrapper[4959]: I1003 15:08:31.451803 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:31 crc kubenswrapper[4959]: I1003 15:08:31.480926 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" podStartSLOduration=3.480907598 podStartE2EDuration="3.480907598s" podCreationTimestamp="2025-10-03 15:08:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:08:31.472939013 +0000 UTC m=+5880.676282440" watchObservedRunningTime="2025-10-03 15:08:31.480907598 +0000 UTC m=+5880.684251015" Oct 03 15:08:38 crc kubenswrapper[4959]: I1003 15:08:38.686058 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:08:38 crc kubenswrapper[4959]: E1003 15:08:38.687117 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:08:38 crc kubenswrapper[4959]: I1003 15:08:38.980344 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.077406 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.077960 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="dnsmasq-dns" containerID="cri-o://2c5ca118997f0d8311b90f962b16584b4c0328746d303bb7a0792d4dd3954ed1" gracePeriod=10 Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.549998 4959 generic.go:334] "Generic (PLEG): container finished" podID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerID="2c5ca118997f0d8311b90f962b16584b4c0328746d303bb7a0792d4dd3954ed1" exitCode=0 Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.550041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" event={"ID":"d3ece186-40f8-4d71-afe1-d2fc1653bd5a","Type":"ContainerDied","Data":"2c5ca118997f0d8311b90f962b16584b4c0328746d303bb7a0792d4dd3954ed1"} Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.706302 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.877906 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config\") pod \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.878010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb\") pod \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.878110 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc\") pod \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.878137 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fphk\" (UniqueName: \"kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk\") pod \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.878250 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb\") pod \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\" (UID: \"d3ece186-40f8-4d71-afe1-d2fc1653bd5a\") " Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.884309 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk" (OuterVolumeSpecName: "kube-api-access-4fphk") pod "d3ece186-40f8-4d71-afe1-d2fc1653bd5a" (UID: "d3ece186-40f8-4d71-afe1-d2fc1653bd5a"). InnerVolumeSpecName "kube-api-access-4fphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.930046 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3ece186-40f8-4d71-afe1-d2fc1653bd5a" (UID: "d3ece186-40f8-4d71-afe1-d2fc1653bd5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.970437 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config" (OuterVolumeSpecName: "config") pod "d3ece186-40f8-4d71-afe1-d2fc1653bd5a" (UID: "d3ece186-40f8-4d71-afe1-d2fc1653bd5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.970954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3ece186-40f8-4d71-afe1-d2fc1653bd5a" (UID: "d3ece186-40f8-4d71-afe1-d2fc1653bd5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.971670 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3ece186-40f8-4d71-afe1-d2fc1653bd5a" (UID: "d3ece186-40f8-4d71-afe1-d2fc1653bd5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.980472 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.980498 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.980509 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.980519 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fphk\" (UniqueName: \"kubernetes.io/projected/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-kube-api-access-4fphk\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:39 crc kubenswrapper[4959]: I1003 15:08:39.980528 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3ece186-40f8-4d71-afe1-d2fc1653bd5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.559899 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" event={"ID":"d3ece186-40f8-4d71-afe1-d2fc1653bd5a","Type":"ContainerDied","Data":"24e3c74c226883cfa3408797d7398517c4106f9675c6de252eba2c3899effefe"} Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.559957 4959 scope.go:117] "RemoveContainer" containerID="2c5ca118997f0d8311b90f962b16584b4c0328746d303bb7a0792d4dd3954ed1" Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.559978 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-696f9966c7-bfs89" Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.585636 4959 scope.go:117] "RemoveContainer" containerID="fa4e295f9888994678dfef497d1bcf1066db6e8a766b755ba7ad12f30e386db9" Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.618097 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:08:40 crc kubenswrapper[4959]: I1003 15:08:40.631029 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-696f9966c7-bfs89"] Oct 03 15:08:41 crc kubenswrapper[4959]: I1003 15:08:41.699005 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" path="/var/lib/kubelet/pods/d3ece186-40f8-4d71-afe1-d2fc1653bd5a/volumes" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.804097 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-thcmv"] Oct 03 15:08:42 crc kubenswrapper[4959]: E1003 15:08:42.804489 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="dnsmasq-dns" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.804502 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="dnsmasq-dns" Oct 03 15:08:42 crc kubenswrapper[4959]: E1003 15:08:42.804527 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="init" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.804533 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="init" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.804732 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ece186-40f8-4d71-afe1-d2fc1653bd5a" containerName="dnsmasq-dns" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.805422 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.837870 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-thcmv"] Oct 03 15:08:42 crc kubenswrapper[4959]: I1003 15:08:42.938777 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xndp2\" (UniqueName: \"kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2\") pod \"cinder-db-create-thcmv\" (UID: \"f90f86b9-e22c-4f37-b44d-774879072a1c\") " pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:43 crc kubenswrapper[4959]: I1003 15:08:43.040516 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xndp2\" (UniqueName: \"kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2\") pod \"cinder-db-create-thcmv\" (UID: \"f90f86b9-e22c-4f37-b44d-774879072a1c\") " pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:43 crc kubenswrapper[4959]: I1003 15:08:43.069605 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xndp2\" (UniqueName: \"kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2\") pod \"cinder-db-create-thcmv\" (UID: \"f90f86b9-e22c-4f37-b44d-774879072a1c\") " pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:43 crc kubenswrapper[4959]: I1003 15:08:43.139598 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:43 crc kubenswrapper[4959]: I1003 15:08:43.624813 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-thcmv"] Oct 03 15:08:43 crc kubenswrapper[4959]: W1003 15:08:43.625825 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf90f86b9_e22c_4f37_b44d_774879072a1c.slice/crio-84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b WatchSource:0}: Error finding container 84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b: Status 404 returned error can't find the container with id 84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b Oct 03 15:08:44 crc kubenswrapper[4959]: I1003 15:08:44.600683 4959 generic.go:334] "Generic (PLEG): container finished" podID="f90f86b9-e22c-4f37-b44d-774879072a1c" containerID="7ab40e3fdd0e277abc0eb32fcca5c08831d7140587aa96478902a76e532be5fb" exitCode=0 Oct 03 15:08:44 crc kubenswrapper[4959]: I1003 15:08:44.600886 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-thcmv" event={"ID":"f90f86b9-e22c-4f37-b44d-774879072a1c","Type":"ContainerDied","Data":"7ab40e3fdd0e277abc0eb32fcca5c08831d7140587aa96478902a76e532be5fb"} Oct 03 15:08:44 crc kubenswrapper[4959]: I1003 15:08:44.600971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-thcmv" event={"ID":"f90f86b9-e22c-4f37-b44d-774879072a1c","Type":"ContainerStarted","Data":"84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b"} Oct 03 15:08:45 crc kubenswrapper[4959]: I1003 15:08:45.989820 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.112458 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xndp2\" (UniqueName: \"kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2\") pod \"f90f86b9-e22c-4f37-b44d-774879072a1c\" (UID: \"f90f86b9-e22c-4f37-b44d-774879072a1c\") " Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.118403 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2" (OuterVolumeSpecName: "kube-api-access-xndp2") pod "f90f86b9-e22c-4f37-b44d-774879072a1c" (UID: "f90f86b9-e22c-4f37-b44d-774879072a1c"). InnerVolumeSpecName "kube-api-access-xndp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.214674 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xndp2\" (UniqueName: \"kubernetes.io/projected/f90f86b9-e22c-4f37-b44d-774879072a1c-kube-api-access-xndp2\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.634256 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-thcmv" Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.634265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-thcmv" event={"ID":"f90f86b9-e22c-4f37-b44d-774879072a1c","Type":"ContainerDied","Data":"84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b"} Oct 03 15:08:46 crc kubenswrapper[4959]: I1003 15:08:46.634309 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84c5fb7080a41bbd159c4b9fafb0e3b6ba471d088fe5af7e333e2f7fe2c8531b" Oct 03 15:08:49 crc kubenswrapper[4959]: I1003 15:08:49.685937 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:08:49 crc kubenswrapper[4959]: E1003 15:08:49.686823 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.956430 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9d9b-account-create-bcdpk"] Oct 03 15:08:52 crc kubenswrapper[4959]: E1003 15:08:52.957110 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90f86b9-e22c-4f37-b44d-774879072a1c" containerName="mariadb-database-create" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.957127 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90f86b9-e22c-4f37-b44d-774879072a1c" containerName="mariadb-database-create" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.957379 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90f86b9-e22c-4f37-b44d-774879072a1c" containerName="mariadb-database-create" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.958030 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.963152 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 03 15:08:52 crc kubenswrapper[4959]: I1003 15:08:52.972617 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9d9b-account-create-bcdpk"] Oct 03 15:08:53 crc kubenswrapper[4959]: I1003 15:08:53.046181 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glqhv\" (UniqueName: \"kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv\") pod \"cinder-9d9b-account-create-bcdpk\" (UID: \"1f8f9179-bec9-4149-8301-7ae6097d8153\") " pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:53 crc kubenswrapper[4959]: I1003 15:08:53.147835 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glqhv\" (UniqueName: \"kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv\") pod \"cinder-9d9b-account-create-bcdpk\" (UID: \"1f8f9179-bec9-4149-8301-7ae6097d8153\") " pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:53 crc kubenswrapper[4959]: I1003 15:08:53.181277 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glqhv\" (UniqueName: \"kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv\") pod \"cinder-9d9b-account-create-bcdpk\" (UID: \"1f8f9179-bec9-4149-8301-7ae6097d8153\") " pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:53 crc kubenswrapper[4959]: I1003 15:08:53.287636 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:53 crc kubenswrapper[4959]: I1003 15:08:53.770484 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9d9b-account-create-bcdpk"] Oct 03 15:08:54 crc kubenswrapper[4959]: I1003 15:08:54.748373 4959 generic.go:334] "Generic (PLEG): container finished" podID="1f8f9179-bec9-4149-8301-7ae6097d8153" containerID="b0f425f4be4223ff24761ccf089044aa8b86b954982ffe4ca020bd410a614b61" exitCode=0 Oct 03 15:08:54 crc kubenswrapper[4959]: I1003 15:08:54.748410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9d9b-account-create-bcdpk" event={"ID":"1f8f9179-bec9-4149-8301-7ae6097d8153","Type":"ContainerDied","Data":"b0f425f4be4223ff24761ccf089044aa8b86b954982ffe4ca020bd410a614b61"} Oct 03 15:08:54 crc kubenswrapper[4959]: I1003 15:08:54.749027 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9d9b-account-create-bcdpk" event={"ID":"1f8f9179-bec9-4149-8301-7ae6097d8153","Type":"ContainerStarted","Data":"d459652eb79c5f4e6a44fee1a4a9b10526c9066aa6f1c901961cc774f1e56d3c"} Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.069313 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.220992 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glqhv\" (UniqueName: \"kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv\") pod \"1f8f9179-bec9-4149-8301-7ae6097d8153\" (UID: \"1f8f9179-bec9-4149-8301-7ae6097d8153\") " Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.227004 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv" (OuterVolumeSpecName: "kube-api-access-glqhv") pod "1f8f9179-bec9-4149-8301-7ae6097d8153" (UID: "1f8f9179-bec9-4149-8301-7ae6097d8153"). InnerVolumeSpecName "kube-api-access-glqhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.323838 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glqhv\" (UniqueName: \"kubernetes.io/projected/1f8f9179-bec9-4149-8301-7ae6097d8153-kube-api-access-glqhv\") on node \"crc\" DevicePath \"\"" Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.773274 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9d9b-account-create-bcdpk" event={"ID":"1f8f9179-bec9-4149-8301-7ae6097d8153","Type":"ContainerDied","Data":"d459652eb79c5f4e6a44fee1a4a9b10526c9066aa6f1c901961cc774f1e56d3c"} Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.773565 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d459652eb79c5f4e6a44fee1a4a9b10526c9066aa6f1c901961cc774f1e56d3c" Oct 03 15:08:56 crc kubenswrapper[4959]: I1003 15:08:56.773339 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9d9b-account-create-bcdpk" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.208929 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-ztg4p"] Oct 03 15:08:58 crc kubenswrapper[4959]: E1003 15:08:58.209525 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8f9179-bec9-4149-8301-7ae6097d8153" containerName="mariadb-account-create" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.209539 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8f9179-bec9-4149-8301-7ae6097d8153" containerName="mariadb-account-create" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.209722 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f8f9179-bec9-4149-8301-7ae6097d8153" containerName="mariadb-account-create" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.210579 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.212621 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xvdm7" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.212879 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.215054 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.224453 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ztg4p"] Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258566 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258693 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.258981 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd4pm\" (UniqueName: \"kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361414 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd4pm\" (UniqueName: \"kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361510 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361585 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.361615 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.365417 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.366425 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.366918 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.367610 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.381279 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.394677 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd4pm\" (UniqueName: \"kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm\") pod \"cinder-db-sync-ztg4p\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:58 crc kubenswrapper[4959]: I1003 15:08:58.544990 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:08:59 crc kubenswrapper[4959]: I1003 15:08:59.028490 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ztg4p"] Oct 03 15:08:59 crc kubenswrapper[4959]: W1003 15:08:59.034930 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aa49515_5afb_4d44_9aef_b5f8b0602773.slice/crio-26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8 WatchSource:0}: Error finding container 26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8: Status 404 returned error can't find the container with id 26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8 Oct 03 15:08:59 crc kubenswrapper[4959]: I1003 15:08:59.802178 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ztg4p" event={"ID":"1aa49515-5afb-4d44-9aef-b5f8b0602773","Type":"ContainerStarted","Data":"99a8906059e98f47b01e40bc5bc64c175edb24eef139f2c4a27eee3bdaa64f90"} Oct 03 15:08:59 crc kubenswrapper[4959]: I1003 15:08:59.802480 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ztg4p" event={"ID":"1aa49515-5afb-4d44-9aef-b5f8b0602773","Type":"ContainerStarted","Data":"26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8"} Oct 03 15:08:59 crc kubenswrapper[4959]: I1003 15:08:59.824266 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-ztg4p" podStartSLOduration=1.824243509 podStartE2EDuration="1.824243509s" podCreationTimestamp="2025-10-03 15:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:08:59.816864048 +0000 UTC m=+5909.020207465" watchObservedRunningTime="2025-10-03 15:08:59.824243509 +0000 UTC m=+5909.027586936" Oct 03 15:09:02 crc kubenswrapper[4959]: I1003 15:09:02.685913 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:09:02 crc kubenswrapper[4959]: E1003 15:09:02.686622 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:09:02 crc kubenswrapper[4959]: I1003 15:09:02.836309 4959 generic.go:334] "Generic (PLEG): container finished" podID="1aa49515-5afb-4d44-9aef-b5f8b0602773" containerID="99a8906059e98f47b01e40bc5bc64c175edb24eef139f2c4a27eee3bdaa64f90" exitCode=0 Oct 03 15:09:02 crc kubenswrapper[4959]: I1003 15:09:02.836365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ztg4p" event={"ID":"1aa49515-5afb-4d44-9aef-b5f8b0602773","Type":"ContainerDied","Data":"99a8906059e98f47b01e40bc5bc64c175edb24eef139f2c4a27eee3bdaa64f90"} Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.196227 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280064 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280131 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280248 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280290 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280443 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd4pm\" (UniqueName: \"kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280532 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data\") pod \"1aa49515-5afb-4d44-9aef-b5f8b0602773\" (UID: \"1aa49515-5afb-4d44-9aef-b5f8b0602773\") " Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.280523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.281056 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1aa49515-5afb-4d44-9aef-b5f8b0602773-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.285540 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.285699 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts" (OuterVolumeSpecName: "scripts") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.286003 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm" (OuterVolumeSpecName: "kube-api-access-rd4pm") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "kube-api-access-rd4pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.304210 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.323579 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data" (OuterVolumeSpecName: "config-data") pod "1aa49515-5afb-4d44-9aef-b5f8b0602773" (UID: "1aa49515-5afb-4d44-9aef-b5f8b0602773"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.382748 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd4pm\" (UniqueName: \"kubernetes.io/projected/1aa49515-5afb-4d44-9aef-b5f8b0602773-kube-api-access-rd4pm\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.382784 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.382794 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.382806 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.382815 4959 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1aa49515-5afb-4d44-9aef-b5f8b0602773-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.855341 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ztg4p" event={"ID":"1aa49515-5afb-4d44-9aef-b5f8b0602773","Type":"ContainerDied","Data":"26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8"} Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.855391 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ztg4p" Oct 03 15:09:04 crc kubenswrapper[4959]: I1003 15:09:04.855403 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26ba8a90476bf1aa4b99624eb9f04c716f9d0bf20f58a5b82bf806aed9faa8f8" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.219308 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:09:05 crc kubenswrapper[4959]: E1003 15:09:05.219883 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa49515-5afb-4d44-9aef-b5f8b0602773" containerName="cinder-db-sync" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.219895 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa49515-5afb-4d44-9aef-b5f8b0602773" containerName="cinder-db-sync" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.220089 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa49515-5afb-4d44-9aef-b5f8b0602773" containerName="cinder-db-sync" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.221139 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.234307 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.306616 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.306679 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.306701 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.306729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s74t4\" (UniqueName: \"kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.306984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.362544 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.364032 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.366943 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xvdm7" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.367172 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.367386 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.367533 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.380634 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410288 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410355 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410554 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s74t4\" (UniqueName: \"kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410632 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410717 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gpgw\" (UniqueName: \"kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410854 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.410966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.411009 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.411101 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.432631 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.433318 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.433889 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.434926 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.466246 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s74t4\" (UniqueName: \"kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4\") pod \"dnsmasq-dns-58dc68b6c7-wb7jn\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.513591 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.513906 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.513954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gpgw\" (UniqueName: \"kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.513995 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.514018 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.514072 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.514113 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.514587 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.514642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.517660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.518179 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.519169 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.528394 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.532362 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gpgw\" (UniqueName: \"kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw\") pod \"cinder-api-0\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " pod="openstack/cinder-api-0" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.544426 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:05 crc kubenswrapper[4959]: I1003 15:09:05.682372 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.021592 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:09:06 crc kubenswrapper[4959]: W1003 15:09:06.030398 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e48b9f_e4d8_49d7_ab1c_e58a62598ece.slice/crio-121dfff4b40b9d4b73c9273c05389190797f8ce0460190df6344ea105f40ab1d WatchSource:0}: Error finding container 121dfff4b40b9d4b73c9273c05389190797f8ce0460190df6344ea105f40ab1d: Status 404 returned error can't find the container with id 121dfff4b40b9d4b73c9273c05389190797f8ce0460190df6344ea105f40ab1d Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.194441 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:06 crc kubenswrapper[4959]: W1003 15:09:06.194447 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc127260_f1e9_448f_aa89_8f141fc0d927.slice/crio-07e38c7045a96f0d5f38229f3ea4e2edce318a7e8919cc6997120d2b9de2533d WatchSource:0}: Error finding container 07e38c7045a96f0d5f38229f3ea4e2edce318a7e8919cc6997120d2b9de2533d: Status 404 returned error can't find the container with id 07e38c7045a96f0d5f38229f3ea4e2edce318a7e8919cc6997120d2b9de2533d Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.874937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerStarted","Data":"96bb4a46a04836ffa6cf9033115d39f2f091bc8e906cb8a8b40d5997ad7a8093"} Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.875273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerStarted","Data":"07e38c7045a96f0d5f38229f3ea4e2edce318a7e8919cc6997120d2b9de2533d"} Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.877326 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerID="ebd5639345efb2019d32c5f977f74c267a945b720c6a74ce75376c696f03e792" exitCode=0 Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.877387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" event={"ID":"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece","Type":"ContainerDied","Data":"ebd5639345efb2019d32c5f977f74c267a945b720c6a74ce75376c696f03e792"} Oct 03 15:09:06 crc kubenswrapper[4959]: I1003 15:09:06.877419 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" event={"ID":"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece","Type":"ContainerStarted","Data":"121dfff4b40b9d4b73c9273c05389190797f8ce0460190df6344ea105f40ab1d"} Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.906168 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerStarted","Data":"b09c3e8db4ebf7cd97490d26e218b883286a7521084748183459aacef13b5c0b"} Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.907112 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.907807 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" event={"ID":"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece","Type":"ContainerStarted","Data":"26a1934fc2088528fd226682b170dd222603ac5c4842aa13d2c59c844fc2ff15"} Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.907912 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.929825 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.929805988 podStartE2EDuration="2.929805988s" podCreationTimestamp="2025-10-03 15:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:07.922866428 +0000 UTC m=+5917.126209855" watchObservedRunningTime="2025-10-03 15:09:07.929805988 +0000 UTC m=+5917.133149405" Oct 03 15:09:07 crc kubenswrapper[4959]: I1003 15:09:07.953790 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" podStartSLOduration=2.953768784 podStartE2EDuration="2.953768784s" podCreationTimestamp="2025-10-03 15:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:07.949648583 +0000 UTC m=+5917.152992000" watchObservedRunningTime="2025-10-03 15:09:07.953768784 +0000 UTC m=+5917.157112201" Oct 03 15:09:15 crc kubenswrapper[4959]: I1003 15:09:15.547480 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:09:15 crc kubenswrapper[4959]: I1003 15:09:15.622754 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:09:15 crc kubenswrapper[4959]: I1003 15:09:15.623213 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="dnsmasq-dns" containerID="cri-o://5a08731f8653d814f3ddd228af571530adba31c6da26e58ae4683fab0451bc7e" gracePeriod=10 Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.019026 4959 generic.go:334] "Generic (PLEG): container finished" podID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerID="5a08731f8653d814f3ddd228af571530adba31c6da26e58ae4683fab0451bc7e" exitCode=0 Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.019247 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" event={"ID":"6c6c71fd-235e-47d6-8220-bfbd1683d915","Type":"ContainerDied","Data":"5a08731f8653d814f3ddd228af571530adba31c6da26e58ae4683fab0451bc7e"} Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.104637 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.202921 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config\") pod \"6c6c71fd-235e-47d6-8220-bfbd1683d915\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.203146 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbt5k\" (UniqueName: \"kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k\") pod \"6c6c71fd-235e-47d6-8220-bfbd1683d915\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.203899 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc\") pod \"6c6c71fd-235e-47d6-8220-bfbd1683d915\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.203929 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb\") pod \"6c6c71fd-235e-47d6-8220-bfbd1683d915\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.204025 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb\") pod \"6c6c71fd-235e-47d6-8220-bfbd1683d915\" (UID: \"6c6c71fd-235e-47d6-8220-bfbd1683d915\") " Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.208954 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k" (OuterVolumeSpecName: "kube-api-access-bbt5k") pod "6c6c71fd-235e-47d6-8220-bfbd1683d915" (UID: "6c6c71fd-235e-47d6-8220-bfbd1683d915"). InnerVolumeSpecName "kube-api-access-bbt5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.246856 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config" (OuterVolumeSpecName: "config") pod "6c6c71fd-235e-47d6-8220-bfbd1683d915" (UID: "6c6c71fd-235e-47d6-8220-bfbd1683d915"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.251169 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c6c71fd-235e-47d6-8220-bfbd1683d915" (UID: "6c6c71fd-235e-47d6-8220-bfbd1683d915"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.265986 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6c6c71fd-235e-47d6-8220-bfbd1683d915" (UID: "6c6c71fd-235e-47d6-8220-bfbd1683d915"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.276251 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6c6c71fd-235e-47d6-8220-bfbd1683d915" (UID: "6c6c71fd-235e-47d6-8220-bfbd1683d915"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.306092 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.306130 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.306144 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.306155 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c6c71fd-235e-47d6-8220-bfbd1683d915-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.306167 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbt5k\" (UniqueName: \"kubernetes.io/projected/6c6c71fd-235e-47d6-8220-bfbd1683d915-kube-api-access-bbt5k\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:16 crc kubenswrapper[4959]: I1003 15:09:16.685670 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:09:16 crc kubenswrapper[4959]: E1003 15:09:16.686221 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.031579 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" event={"ID":"6c6c71fd-235e-47d6-8220-bfbd1683d915","Type":"ContainerDied","Data":"a8c87eae8e69d30268ccd218465ab1f739e41e79a259eb3ef7b8eaf446b578db"} Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.031921 4959 scope.go:117] "RemoveContainer" containerID="5a08731f8653d814f3ddd228af571530adba31c6da26e58ae4683fab0451bc7e" Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.032079 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64b8d7d4fc-ggxmr" Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.067755 4959 scope.go:117] "RemoveContainer" containerID="ef0e6ba0e15f69530d4ca8bde7b3b2664560364a20f32f7a16de743c888d6169" Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.079368 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.086322 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64b8d7d4fc-ggxmr"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.305584 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.305824 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cd456263-94c7-4963-97a4-fbaa389c2a9c" containerName="nova-scheduler-scheduler" containerID="cri-o://d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.320561 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.320850 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" containerID="cri-o://941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.321186 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" containerID="cri-o://70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.341055 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.341299 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-log" containerID="cri-o://12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.343058 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-api" containerID="cri-o://0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.357690 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.357927 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerName="nova-cell1-conductor-conductor" containerID="cri-o://310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.372710 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.372985 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4dcab959-eef4-4d50-9d7e-1ae36849c889" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://011b44e5eb3ea9ec941ba1fa482024d8835f9b68364c5eddd0ba0d656936540f" gracePeriod=30 Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.548839 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 03 15:09:17 crc kubenswrapper[4959]: I1003 15:09:17.702245 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" path="/var/lib/kubelet/pods/6c6c71fd-235e-47d6-8220-bfbd1683d915/volumes" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.049524 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d31aaca-d388-4930-9efe-35718e9898f5" containerID="941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0" exitCode=143 Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.049584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerDied","Data":"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0"} Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.051140 4959 generic.go:334] "Generic (PLEG): container finished" podID="4dcab959-eef4-4d50-9d7e-1ae36849c889" containerID="011b44e5eb3ea9ec941ba1fa482024d8835f9b68364c5eddd0ba0d656936540f" exitCode=0 Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.051216 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4dcab959-eef4-4d50-9d7e-1ae36849c889","Type":"ContainerDied","Data":"011b44e5eb3ea9ec941ba1fa482024d8835f9b68364c5eddd0ba0d656936540f"} Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.053690 4959 generic.go:334] "Generic (PLEG): container finished" podID="472fb0da-e872-442c-b2ce-953454e5d61a" containerID="12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955" exitCode=143 Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.053738 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerDied","Data":"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955"} Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.218208 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.345124 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data\") pod \"4dcab959-eef4-4d50-9d7e-1ae36849c889\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.346539 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle\") pod \"4dcab959-eef4-4d50-9d7e-1ae36849c889\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.346669 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmb87\" (UniqueName: \"kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87\") pod \"4dcab959-eef4-4d50-9d7e-1ae36849c889\" (UID: \"4dcab959-eef4-4d50-9d7e-1ae36849c889\") " Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.356673 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87" (OuterVolumeSpecName: "kube-api-access-pmb87") pod "4dcab959-eef4-4d50-9d7e-1ae36849c889" (UID: "4dcab959-eef4-4d50-9d7e-1ae36849c889"). InnerVolumeSpecName "kube-api-access-pmb87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.381276 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data" (OuterVolumeSpecName: "config-data") pod "4dcab959-eef4-4d50-9d7e-1ae36849c889" (UID: "4dcab959-eef4-4d50-9d7e-1ae36849c889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.382545 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4dcab959-eef4-4d50-9d7e-1ae36849c889" (UID: "4dcab959-eef4-4d50-9d7e-1ae36849c889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.456139 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.456573 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmb87\" (UniqueName: \"kubernetes.io/projected/4dcab959-eef4-4d50-9d7e-1ae36849c889-kube-api-access-pmb87\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:18 crc kubenswrapper[4959]: I1003 15:09:18.456666 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4dcab959-eef4-4d50-9d7e-1ae36849c889-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.065148 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4dcab959-eef4-4d50-9d7e-1ae36849c889","Type":"ContainerDied","Data":"3d0187dafa2551ab860390b927b518e3434c66fe04e99e921c7afdf0fda356de"} Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.065255 4959 scope.go:117] "RemoveContainer" containerID="011b44e5eb3ea9ec941ba1fa482024d8835f9b68364c5eddd0ba0d656936540f" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.065285 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.115848 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.123593 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.134259 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:19 crc kubenswrapper[4959]: E1003 15:09:19.135027 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dcab959-eef4-4d50-9d7e-1ae36849c889" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.135134 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dcab959-eef4-4d50-9d7e-1ae36849c889" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 15:09:19 crc kubenswrapper[4959]: E1003 15:09:19.135263 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="init" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.135345 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="init" Oct 03 15:09:19 crc kubenswrapper[4959]: E1003 15:09:19.135456 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="dnsmasq-dns" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.135536 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="dnsmasq-dns" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.135856 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6c71fd-235e-47d6-8220-bfbd1683d915" containerName="dnsmasq-dns" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.135971 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dcab959-eef4-4d50-9d7e-1ae36849c889" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.136916 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.140383 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.141815 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.170247 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld9qd\" (UniqueName: \"kubernetes.io/projected/84757a0b-ee0a-443c-8b44-b3aeaae6785e-kube-api-access-ld9qd\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.170405 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.170492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.272444 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld9qd\" (UniqueName: \"kubernetes.io/projected/84757a0b-ee0a-443c-8b44-b3aeaae6785e-kube-api-access-ld9qd\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.272814 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.272989 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.278817 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.283732 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84757a0b-ee0a-443c-8b44-b3aeaae6785e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.289329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld9qd\" (UniqueName: \"kubernetes.io/projected/84757a0b-ee0a-443c-8b44-b3aeaae6785e-kube-api-access-ld9qd\") pod \"nova-cell1-novncproxy-0\" (UID: \"84757a0b-ee0a-443c-8b44-b3aeaae6785e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.451587 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.700212 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dcab959-eef4-4d50-9d7e-1ae36849c889" path="/var/lib/kubelet/pods/4dcab959-eef4-4d50-9d7e-1ae36849c889/volumes" Oct 03 15:09:19 crc kubenswrapper[4959]: I1003 15:09:19.970349 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.094378 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"84757a0b-ee0a-443c-8b44-b3aeaae6785e","Type":"ContainerStarted","Data":"26e8bd9e8db83373c836f61e60c857b4d07b53626c3f75ec2c58d7987bf675f3"} Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.463528 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": read tcp 10.217.0.2:56808->10.217.1.76:8775: read: connection reset by peer" Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.465357 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": read tcp 10.217.0.2:56820->10.217.1.76:8775: read: connection reset by peer" Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.577283 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.577539 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f" gracePeriod=30 Oct 03 15:09:20 crc kubenswrapper[4959]: I1003 15:09:20.857827 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.001304 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.010356 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d62wj\" (UniqueName: \"kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj\") pod \"cd456263-94c7-4963-97a4-fbaa389c2a9c\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.010421 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle\") pod \"cd456263-94c7-4963-97a4-fbaa389c2a9c\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.010516 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data\") pod \"cd456263-94c7-4963-97a4-fbaa389c2a9c\" (UID: \"cd456263-94c7-4963-97a4-fbaa389c2a9c\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.016948 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj" (OuterVolumeSpecName: "kube-api-access-d62wj") pod "cd456263-94c7-4963-97a4-fbaa389c2a9c" (UID: "cd456263-94c7-4963-97a4-fbaa389c2a9c"). InnerVolumeSpecName "kube-api-access-d62wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.061279 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data" (OuterVolumeSpecName: "config-data") pod "cd456263-94c7-4963-97a4-fbaa389c2a9c" (UID: "cd456263-94c7-4963-97a4-fbaa389c2a9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.063313 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd456263-94c7-4963-97a4-fbaa389c2a9c" (UID: "cd456263-94c7-4963-97a4-fbaa389c2a9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.082188 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.112172 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndfbv\" (UniqueName: \"kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv\") pod \"4d31aaca-d388-4930-9efe-35718e9898f5\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.113464 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data\") pod \"4d31aaca-d388-4930-9efe-35718e9898f5\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.113699 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle\") pod \"4d31aaca-d388-4930-9efe-35718e9898f5\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.113802 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs\") pod \"4d31aaca-d388-4930-9efe-35718e9898f5\" (UID: \"4d31aaca-d388-4930-9efe-35718e9898f5\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.114754 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv" (OuterVolumeSpecName: "kube-api-access-ndfbv") pod "4d31aaca-d388-4930-9efe-35718e9898f5" (UID: "4d31aaca-d388-4930-9efe-35718e9898f5"). InnerVolumeSpecName "kube-api-access-ndfbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.115116 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs" (OuterVolumeSpecName: "logs") pod "4d31aaca-d388-4930-9efe-35718e9898f5" (UID: "4d31aaca-d388-4930-9efe-35718e9898f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.116008 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d62wj\" (UniqueName: \"kubernetes.io/projected/cd456263-94c7-4963-97a4-fbaa389c2a9c-kube-api-access-d62wj\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.116092 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.116149 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d31aaca-d388-4930-9efe-35718e9898f5-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.116222 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd456263-94c7-4963-97a4-fbaa389c2a9c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.116277 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndfbv\" (UniqueName: \"kubernetes.io/projected/4d31aaca-d388-4930-9efe-35718e9898f5-kube-api-access-ndfbv\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.146668 4959 generic.go:334] "Generic (PLEG): container finished" podID="472fb0da-e872-442c-b2ce-953454e5d61a" containerID="0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec" exitCode=0 Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.146740 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.146749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerDied","Data":"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.146779 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"472fb0da-e872-442c-b2ce-953454e5d61a","Type":"ContainerDied","Data":"0e7d26d05aa4ec888d5d589ff9c87fccbacd0656094495d10f5c08b808709787"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.146795 4959 scope.go:117] "RemoveContainer" containerID="0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.174340 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data" (OuterVolumeSpecName: "config-data") pod "4d31aaca-d388-4930-9efe-35718e9898f5" (UID: "4d31aaca-d388-4930-9efe-35718e9898f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.174495 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d31aaca-d388-4930-9efe-35718e9898f5" (UID: "4d31aaca-d388-4930-9efe-35718e9898f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.174859 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d31aaca-d388-4930-9efe-35718e9898f5" containerID="70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c" exitCode=0 Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.174998 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.175411 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerDied","Data":"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.175467 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4d31aaca-d388-4930-9efe-35718e9898f5","Type":"ContainerDied","Data":"924f5ec3147ea8dfe588e6a6cfde8f7d7b534f3707aa62c8ed86cc6d018f8960"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.180940 4959 generic.go:334] "Generic (PLEG): container finished" podID="cd456263-94c7-4963-97a4-fbaa389c2a9c" containerID="d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4" exitCode=0 Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.181002 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd456263-94c7-4963-97a4-fbaa389c2a9c","Type":"ContainerDied","Data":"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.181025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd456263-94c7-4963-97a4-fbaa389c2a9c","Type":"ContainerDied","Data":"3b8560c4c4e9c16e45af4c1784b62615951d05f460975f180fe738f82b29a8d1"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.181073 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.188322 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"84757a0b-ee0a-443c-8b44-b3aeaae6785e","Type":"ContainerStarted","Data":"bebd6e860b4e0f258c6545d831c632a2b6c18fb50b32dd808d384201271203df"} Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.198172 4959 scope.go:117] "RemoveContainer" containerID="12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.215312 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.21529559 podStartE2EDuration="2.21529559s" podCreationTimestamp="2025-10-03 15:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:21.205539601 +0000 UTC m=+5930.408883018" watchObservedRunningTime="2025-10-03 15:09:21.21529559 +0000 UTC m=+5930.418639007" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.218244 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data\") pod \"472fb0da-e872-442c-b2ce-953454e5d61a\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.218448 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle\") pod \"472fb0da-e872-442c-b2ce-953454e5d61a\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.218605 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs\") pod \"472fb0da-e872-442c-b2ce-953454e5d61a\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.218687 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs9xx\" (UniqueName: \"kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx\") pod \"472fb0da-e872-442c-b2ce-953454e5d61a\" (UID: \"472fb0da-e872-442c-b2ce-953454e5d61a\") " Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.219180 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.219306 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d31aaca-d388-4930-9efe-35718e9898f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.222672 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs" (OuterVolumeSpecName: "logs") pod "472fb0da-e872-442c-b2ce-953454e5d61a" (UID: "472fb0da-e872-442c-b2ce-953454e5d61a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.229976 4959 scope.go:117] "RemoveContainer" containerID="0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.230494 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec\": container with ID starting with 0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec not found: ID does not exist" containerID="0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.230552 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec"} err="failed to get container status \"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec\": rpc error: code = NotFound desc = could not find container \"0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec\": container with ID starting with 0b1be431f5c419c2329925a5248dc2d1536c323d77df1a5dae28de38ccb22bec not found: ID does not exist" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.230577 4959 scope.go:117] "RemoveContainer" containerID="12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.230814 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx" (OuterVolumeSpecName: "kube-api-access-zs9xx") pod "472fb0da-e872-442c-b2ce-953454e5d61a" (UID: "472fb0da-e872-442c-b2ce-953454e5d61a"). InnerVolumeSpecName "kube-api-access-zs9xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.232850 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955\": container with ID starting with 12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955 not found: ID does not exist" containerID="12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.232880 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955"} err="failed to get container status \"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955\": rpc error: code = NotFound desc = could not find container \"12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955\": container with ID starting with 12f2e28dc610bbb9c60f13d75c96017949308e9ea08b6310b64f4950b2e32955 not found: ID does not exist" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.232900 4959 scope.go:117] "RemoveContainer" containerID="70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.261330 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "472fb0da-e872-442c-b2ce-953454e5d61a" (UID: "472fb0da-e872-442c-b2ce-953454e5d61a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.270326 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.288594 4959 scope.go:117] "RemoveContainer" containerID="941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.322087 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/472fb0da-e872-442c-b2ce-953454e5d61a-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.322121 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs9xx\" (UniqueName: \"kubernetes.io/projected/472fb0da-e872-442c-b2ce-953454e5d61a-kube-api-access-zs9xx\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.322132 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.340275 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.354285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data" (OuterVolumeSpecName: "config-data") pod "472fb0da-e872-442c-b2ce-953454e5d61a" (UID: "472fb0da-e872-442c-b2ce-953454e5d61a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.354678 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.363519 4959 scope.go:117] "RemoveContainer" containerID="70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.369161 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.369574 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c\": container with ID starting with 70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c not found: ID does not exist" containerID="70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.369618 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c"} err="failed to get container status \"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c\": rpc error: code = NotFound desc = could not find container \"70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c\": container with ID starting with 70d43b29873c944202203a4d632753da895a146b8a20e5bdd86b6711068af73c not found: ID does not exist" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.369649 4959 scope.go:117] "RemoveContainer" containerID="941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.370073 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0\": container with ID starting with 941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0 not found: ID does not exist" containerID="941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.370121 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0"} err="failed to get container status \"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0\": rpc error: code = NotFound desc = could not find container \"941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0\": container with ID starting with 941e3a7b6b0574fd32141d0fb09a7c6a76af3773129b4e537b5b321d1f08f4e0 not found: ID does not exist" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.370146 4959 scope.go:117] "RemoveContainer" containerID="d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.383898 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.384493 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd456263-94c7-4963-97a4-fbaa389c2a9c" containerName="nova-scheduler-scheduler" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384553 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd456263-94c7-4963-97a4-fbaa389c2a9c" containerName="nova-scheduler-scheduler" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.384569 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384579 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.384597 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384606 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.384643 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-log" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384655 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-log" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.384671 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-api" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384680 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-api" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384901 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd456263-94c7-4963-97a4-fbaa389c2a9c" containerName="nova-scheduler-scheduler" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384926 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-log" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384942 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-metadata" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384965 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" containerName="nova-api-api" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.384982 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" containerName="nova-metadata-log" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.385809 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.388428 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.391483 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.403314 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.404828 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.408222 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.411497 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.436447 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/472fb0da-e872-442c-b2ce-953454e5d61a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.438953 4959 scope.go:117] "RemoveContainer" containerID="d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4" Oct 03 15:09:21 crc kubenswrapper[4959]: E1003 15:09:21.439606 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4\": container with ID starting with d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4 not found: ID does not exist" containerID="d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.439634 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4"} err="failed to get container status \"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4\": rpc error: code = NotFound desc = could not find container \"d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4\": container with ID starting with d2351a61d40ed8791f0f2427cd569ba900b7cec8b9ce29356dd1179a4c365ad4 not found: ID does not exist" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.490548 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.503639 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.510933 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.512793 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.515416 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.533878 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538399 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9jk6\" (UniqueName: \"kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538441 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538531 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdrtx\" (UniqueName: \"kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538548 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538581 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.538624 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.639773 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.639835 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.639882 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.639905 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.639971 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640138 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640369 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640381 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9jk6\" (UniqueName: \"kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640553 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640593 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz6ql\" (UniqueName: \"kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640815 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdrtx\" (UniqueName: \"kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.640850 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.645285 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.646229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.652076 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.653105 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.657504 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9jk6\" (UniqueName: \"kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6\") pod \"nova-scheduler-0\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.663121 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdrtx\" (UniqueName: \"kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx\") pod \"nova-metadata-0\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.699907 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="472fb0da-e872-442c-b2ce-953454e5d61a" path="/var/lib/kubelet/pods/472fb0da-e872-442c-b2ce-953454e5d61a/volumes" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.700535 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d31aaca-d388-4930-9efe-35718e9898f5" path="/var/lib/kubelet/pods/4d31aaca-d388-4930-9efe-35718e9898f5/volumes" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.701280 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd456263-94c7-4963-97a4-fbaa389c2a9c" path="/var/lib/kubelet/pods/cd456263-94c7-4963-97a4-fbaa389c2a9c/volumes" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.743356 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.743476 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.743535 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.743607 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz6ql\" (UniqueName: \"kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.744229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.749759 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.750315 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.751968 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.762502 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.763928 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz6ql\" (UniqueName: \"kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql\") pod \"nova-api-0\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " pod="openstack/nova-api-0" Oct 03 15:09:21 crc kubenswrapper[4959]: I1003 15:09:21.835791 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 15:09:22 crc kubenswrapper[4959]: I1003 15:09:22.269096 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 15:09:22 crc kubenswrapper[4959]: E1003 15:09:22.307265 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 15:09:22 crc kubenswrapper[4959]: E1003 15:09:22.308721 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 15:09:22 crc kubenswrapper[4959]: E1003 15:09:22.312821 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 15:09:22 crc kubenswrapper[4959]: E1003 15:09:22.312893 4959 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerName="nova-cell1-conductor-conductor" Oct 03 15:09:22 crc kubenswrapper[4959]: I1003 15:09:22.334918 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 15:09:22 crc kubenswrapper[4959]: I1003 15:09:22.399971 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.193573 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.217870 4959 generic.go:334] "Generic (PLEG): container finished" podID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" containerID="d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f" exitCode=0 Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.218165 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4d460ff8-f592-4b23-9650-e0f08fa3daa3","Type":"ContainerDied","Data":"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.218207 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4d460ff8-f592-4b23-9650-e0f08fa3daa3","Type":"ContainerDied","Data":"2edea1e3bcfdb47e75f51085e8af5efc901a0fb7612e3e6189d54984ca92ea80"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.218225 4959 scope.go:117] "RemoveContainer" containerID="d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.218315 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.222715 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerStarted","Data":"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.222753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerStarted","Data":"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.222765 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerStarted","Data":"0ed501ea1ef3f8b2623b3ea1992bb9a57c4f79de996bdb9e0dc10b479cd7c23a"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.230646 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91f1ca45-586a-4462-85e2-385dccbcc0c1","Type":"ContainerStarted","Data":"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.230693 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91f1ca45-586a-4462-85e2-385dccbcc0c1","Type":"ContainerStarted","Data":"1d1205c30eb8212e9aa08083d15196b180a7665fc485ea79a098a28afb9ea04f"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.232767 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerStarted","Data":"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.232798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerStarted","Data":"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.232812 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerStarted","Data":"7c7c336e549e676d3de05ff33ea5fb5477883424c37e80ae5f1daf2bf415a5a6"} Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.243262 4959 scope.go:117] "RemoveContainer" containerID="d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f" Oct 03 15:09:23 crc kubenswrapper[4959]: E1003 15:09:23.244673 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f\": container with ID starting with d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f not found: ID does not exist" containerID="d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.244714 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f"} err="failed to get container status \"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f\": rpc error: code = NotFound desc = could not find container \"d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f\": container with ID starting with d405c44c161a151f59ac7f0f2b64c189141e6db11a7119618fd9ec9e1531d64f not found: ID does not exist" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.264655 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.264637311 podStartE2EDuration="2.264637311s" podCreationTimestamp="2025-10-03 15:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:23.248501286 +0000 UTC m=+5932.451844743" watchObservedRunningTime="2025-10-03 15:09:23.264637311 +0000 UTC m=+5932.467980728" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.273627 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data\") pod \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.273776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppwwz\" (UniqueName: \"kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz\") pod \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.273882 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") pod \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.285008 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz" (OuterVolumeSpecName: "kube-api-access-ppwwz") pod "4d460ff8-f592-4b23-9650-e0f08fa3daa3" (UID: "4d460ff8-f592-4b23-9650-e0f08fa3daa3"). InnerVolumeSpecName "kube-api-access-ppwwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.291173 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.291152529 podStartE2EDuration="2.291152529s" podCreationTimestamp="2025-10-03 15:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:23.285548152 +0000 UTC m=+5932.488891569" watchObservedRunningTime="2025-10-03 15:09:23.291152529 +0000 UTC m=+5932.494495946" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.307360 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.307344256 podStartE2EDuration="2.307344256s" podCreationTimestamp="2025-10-03 15:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:23.30464231 +0000 UTC m=+5932.507985727" watchObservedRunningTime="2025-10-03 15:09:23.307344256 +0000 UTC m=+5932.510687673" Oct 03 15:09:23 crc kubenswrapper[4959]: E1003 15:09:23.309724 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle podName:4d460ff8-f592-4b23-9650-e0f08fa3daa3 nodeName:}" failed. No retries permitted until 2025-10-03 15:09:23.809690643 +0000 UTC m=+5933.013034060 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle") pod "4d460ff8-f592-4b23-9650-e0f08fa3daa3" (UID: "4d460ff8-f592-4b23-9650-e0f08fa3daa3") : error deleting /var/lib/kubelet/pods/4d460ff8-f592-4b23-9650-e0f08fa3daa3/volume-subpaths: remove /var/lib/kubelet/pods/4d460ff8-f592-4b23-9650-e0f08fa3daa3/volume-subpaths: no such file or directory Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.312519 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data" (OuterVolumeSpecName: "config-data") pod "4d460ff8-f592-4b23-9650-e0f08fa3daa3" (UID: "4d460ff8-f592-4b23-9650-e0f08fa3daa3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.376138 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppwwz\" (UniqueName: \"kubernetes.io/projected/4d460ff8-f592-4b23-9650-e0f08fa3daa3-kube-api-access-ppwwz\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.376174 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.887267 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") pod \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\" (UID: \"4d460ff8-f592-4b23-9650-e0f08fa3daa3\") " Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.891742 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d460ff8-f592-4b23-9650-e0f08fa3daa3" (UID: "4d460ff8-f592-4b23-9650-e0f08fa3daa3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:23 crc kubenswrapper[4959]: I1003 15:09:23.989334 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d460ff8-f592-4b23-9650-e0f08fa3daa3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.148421 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.156901 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.207376 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:24 crc kubenswrapper[4959]: E1003 15:09:24.207783 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" containerName="nova-cell0-conductor-conductor" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.207802 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" containerName="nova-cell0-conductor-conductor" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.207990 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" containerName="nova-cell0-conductor-conductor" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.208659 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.210708 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.224326 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.300068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.300123 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.300148 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lflfp\" (UniqueName: \"kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.402075 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.402158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.402181 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lflfp\" (UniqueName: \"kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.407855 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.408385 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.418960 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lflfp\" (UniqueName: \"kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp\") pod \"nova-cell0-conductor-0\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.452684 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.524008 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:24 crc kubenswrapper[4959]: I1003 15:09:24.976926 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 15:09:25 crc kubenswrapper[4959]: I1003 15:09:25.258356 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fddc10e1-3a1b-4abf-8277-742f68ab96c2","Type":"ContainerStarted","Data":"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a"} Oct 03 15:09:25 crc kubenswrapper[4959]: I1003 15:09:25.258615 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fddc10e1-3a1b-4abf-8277-742f68ab96c2","Type":"ContainerStarted","Data":"4f5e49c44658be4208aa6851cce254fc23a3d9c54100677fa7d4c715dc102d6d"} Oct 03 15:09:25 crc kubenswrapper[4959]: I1003 15:09:25.258985 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:25 crc kubenswrapper[4959]: I1003 15:09:25.283920 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.283897766 podStartE2EDuration="1.283897766s" podCreationTimestamp="2025-10-03 15:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:25.278113734 +0000 UTC m=+5934.481457141" watchObservedRunningTime="2025-10-03 15:09:25.283897766 +0000 UTC m=+5934.487241183" Oct 03 15:09:25 crc kubenswrapper[4959]: I1003 15:09:25.700829 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d460ff8-f592-4b23-9650-e0f08fa3daa3" path="/var/lib/kubelet/pods/4d460ff8-f592-4b23-9650-e0f08fa3daa3/volumes" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.645236 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.750826 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.751419 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpbhp\" (UniqueName: \"kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp\") pod \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.752398 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle\") pod \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.752626 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data\") pod \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\" (UID: \"55d84893-19fb-4b35-a5fa-8e45af23ec3a\") " Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.757736 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp" (OuterVolumeSpecName: "kube-api-access-vpbhp") pod "55d84893-19fb-4b35-a5fa-8e45af23ec3a" (UID: "55d84893-19fb-4b35-a5fa-8e45af23ec3a"). InnerVolumeSpecName "kube-api-access-vpbhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.763602 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.763768 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.782991 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55d84893-19fb-4b35-a5fa-8e45af23ec3a" (UID: "55d84893-19fb-4b35-a5fa-8e45af23ec3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.784003 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data" (OuterVolumeSpecName: "config-data") pod "55d84893-19fb-4b35-a5fa-8e45af23ec3a" (UID: "55d84893-19fb-4b35-a5fa-8e45af23ec3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.855437 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpbhp\" (UniqueName: \"kubernetes.io/projected/55d84893-19fb-4b35-a5fa-8e45af23ec3a-kube-api-access-vpbhp\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.855467 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:26 crc kubenswrapper[4959]: I1003 15:09:26.855477 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d84893-19fb-4b35-a5fa-8e45af23ec3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.275870 4959 generic.go:334] "Generic (PLEG): container finished" podID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" exitCode=0 Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.275949 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"55d84893-19fb-4b35-a5fa-8e45af23ec3a","Type":"ContainerDied","Data":"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203"} Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.275981 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"55d84893-19fb-4b35-a5fa-8e45af23ec3a","Type":"ContainerDied","Data":"f98fd033b52d882adf827008ebe435d02a74778f4a192b0a39369f4f165ecc88"} Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.276003 4959 scope.go:117] "RemoveContainer" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.276059 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.300098 4959 scope.go:117] "RemoveContainer" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" Oct 03 15:09:27 crc kubenswrapper[4959]: E1003 15:09:27.301701 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203\": container with ID starting with 310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203 not found: ID does not exist" containerID="310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.301779 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203"} err="failed to get container status \"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203\": rpc error: code = NotFound desc = could not find container \"310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203\": container with ID starting with 310b3be7d16ab761f20e4dad265bf09e62b733d9008d1d53519839b5ddcc1203 not found: ID does not exist" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.319433 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.327206 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.336254 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:27 crc kubenswrapper[4959]: E1003 15:09:27.337023 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerName="nova-cell1-conductor-conductor" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.337111 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerName="nova-cell1-conductor-conductor" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.337374 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" containerName="nova-cell1-conductor-conductor" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.338064 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.341546 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.344708 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.467084 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.467152 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7w8b\" (UniqueName: \"kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.467282 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.568875 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.569033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.569101 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7w8b\" (UniqueName: \"kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.573723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.583771 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.616540 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7w8b\" (UniqueName: \"kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b\") pod \"nova-cell1-conductor-0\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.664034 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:27 crc kubenswrapper[4959]: I1003 15:09:27.695784 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55d84893-19fb-4b35-a5fa-8e45af23ec3a" path="/var/lib/kubelet/pods/55d84893-19fb-4b35-a5fa-8e45af23ec3a/volumes" Oct 03 15:09:28 crc kubenswrapper[4959]: I1003 15:09:28.113181 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 15:09:28 crc kubenswrapper[4959]: W1003 15:09:28.116310 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd000b1a0_ae4e_4167_aef4_88cffa9b54cd.slice/crio-00689e23c88b06113f86896cdcad526bf420835a403cebaf25793d6a09167c94 WatchSource:0}: Error finding container 00689e23c88b06113f86896cdcad526bf420835a403cebaf25793d6a09167c94: Status 404 returned error can't find the container with id 00689e23c88b06113f86896cdcad526bf420835a403cebaf25793d6a09167c94 Oct 03 15:09:28 crc kubenswrapper[4959]: I1003 15:09:28.291541 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d000b1a0-ae4e-4167-aef4-88cffa9b54cd","Type":"ContainerStarted","Data":"a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af"} Oct 03 15:09:28 crc kubenswrapper[4959]: I1003 15:09:28.291583 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d000b1a0-ae4e-4167-aef4-88cffa9b54cd","Type":"ContainerStarted","Data":"00689e23c88b06113f86896cdcad526bf420835a403cebaf25793d6a09167c94"} Oct 03 15:09:28 crc kubenswrapper[4959]: I1003 15:09:28.291680 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:29 crc kubenswrapper[4959]: I1003 15:09:29.452730 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:29 crc kubenswrapper[4959]: I1003 15:09:29.474054 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:29 crc kubenswrapper[4959]: I1003 15:09:29.502798 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.502773951 podStartE2EDuration="2.502773951s" podCreationTimestamp="2025-10-03 15:09:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:28.310647536 +0000 UTC m=+5937.513990993" watchObservedRunningTime="2025-10-03 15:09:29.502773951 +0000 UTC m=+5938.706117378" Oct 03 15:09:30 crc kubenswrapper[4959]: I1003 15:09:30.328393 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 03 15:09:30 crc kubenswrapper[4959]: I1003 15:09:30.685133 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:09:30 crc kubenswrapper[4959]: E1003 15:09:30.685413 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.752654 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.765890 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.765926 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.804455 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.840224 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:09:31 crc kubenswrapper[4959]: I1003 15:09:31.840271 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 15:09:32 crc kubenswrapper[4959]: I1003 15:09:32.367505 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 15:09:32 crc kubenswrapper[4959]: I1003 15:09:32.848406 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:09:32 crc kubenswrapper[4959]: I1003 15:09:32.848438 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:09:32 crc kubenswrapper[4959]: I1003 15:09:32.930369 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:09:32 crc kubenswrapper[4959]: I1003 15:09:32.930366 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:09:34 crc kubenswrapper[4959]: I1003 15:09:34.547702 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.079606 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.082346 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.089152 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.108536 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.247386 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.247740 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.247818 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgc8d\" (UniqueName: \"kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.247928 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.248001 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.248098 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350528 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350624 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgc8d\" (UniqueName: \"kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350741 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.350915 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.351108 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.356460 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.361264 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.364774 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.370934 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.380247 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgc8d\" (UniqueName: \"kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d\") pod \"cinder-scheduler-0\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.418869 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:36 crc kubenswrapper[4959]: I1003 15:09:36.840881 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.137423 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.139134 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api-log" containerID="cri-o://96bb4a46a04836ffa6cf9033115d39f2f091bc8e906cb8a8b40d5997ad7a8093" gracePeriod=30 Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.139288 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api" containerID="cri-o://b09c3e8db4ebf7cd97490d26e218b883286a7521084748183459aacef13b5c0b" gracePeriod=30 Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.387954 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerStarted","Data":"0a8b468ae333eb960ef77d53290ab72ee0fe48b5386becb2f828cecc0816a80a"} Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.391848 4959 generic.go:334] "Generic (PLEG): container finished" podID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerID="96bb4a46a04836ffa6cf9033115d39f2f091bc8e906cb8a8b40d5997ad7a8093" exitCode=143 Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.391895 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerDied","Data":"96bb4a46a04836ffa6cf9033115d39f2f091bc8e906cb8a8b40d5997ad7a8093"} Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.716846 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.810033 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.811593 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.814683 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.825355 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981171 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981385 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981450 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981472 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981514 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-run\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981595 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981651 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981680 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981725 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981798 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981817 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981913 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdvtg\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-kube-api-access-mdvtg\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981946 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:37 crc kubenswrapper[4959]: I1003 15:09:37.981977 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.083251 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.083757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.083870 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.083957 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084092 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084423 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084257 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.083917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084211 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084712 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084827 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdvtg\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-kube-api-access-mdvtg\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.084933 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085144 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085302 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085412 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085509 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085604 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-run\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.085742 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-run\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.086057 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.086930 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.086976 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.087081 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.087208 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.087350 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/363bf001-9d71-4656-ab22-9bc29fbb88c3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.087685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.089000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.097402 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.101096 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdvtg\" (UniqueName: \"kubernetes.io/projected/363bf001-9d71-4656-ab22-9bc29fbb88c3-kube-api-access-mdvtg\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.101881 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363bf001-9d71-4656-ab22-9bc29fbb88c3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"363bf001-9d71-4656-ab22-9bc29fbb88c3\") " pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.139797 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.275173 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.278244 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.282107 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.300785 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391633 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5w46\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-kube-api-access-z5w46\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391684 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-dev\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391720 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391745 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391761 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-ceph\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391784 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391810 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391850 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391864 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-sys\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391897 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-lib-modules\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.391914 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.392048 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-run\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.392071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-scripts\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.392088 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.392110 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.405898 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerStarted","Data":"08326680f8abe2848e00a23d38ff42e4293f18ef8c72c26e02af8504afcd4a0a"} Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.405936 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerStarted","Data":"9f0c39d9b0b43ca3320034aee406c57377c48e9e141639250c6d80422cf7c680"} Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.467979 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.467959697 podStartE2EDuration="2.467959697s" podCreationTimestamp="2025-10-03 15:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:38.465455815 +0000 UTC m=+5947.668799232" watchObservedRunningTime="2025-10-03 15:09:38.467959697 +0000 UTC m=+5947.671303104" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493614 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493662 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5w46\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-kube-api-access-z5w46\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493690 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-dev\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493690 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493728 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493786 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-ceph\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493807 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493830 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493889 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493903 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-sys\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-lib-modules\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.493985 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-run\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494003 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-scripts\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494323 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494391 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-dev\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494686 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-sys\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494868 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-lib-modules\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.494956 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.495249 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.495245 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.495386 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.495272 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-run\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.503757 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.504240 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.504353 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-ceph\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.504642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-scripts\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.508762 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.525973 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5w46\" (UniqueName: \"kubernetes.io/projected/c4d9369a-f74d-4c94-8578-41c4cd5beb0d-kube-api-access-z5w46\") pod \"cinder-backup-0\" (UID: \"c4d9369a-f74d-4c94-8578-41c4cd5beb0d\") " pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.605868 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.791461 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 15:09:38 crc kubenswrapper[4959]: W1003 15:09:38.808951 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod363bf001_9d71_4656_ab22_9bc29fbb88c3.slice/crio-06e086f23bcd8914b97606fcca05875573e5433d2dbad23011b81c8c242b8c52 WatchSource:0}: Error finding container 06e086f23bcd8914b97606fcca05875573e5433d2dbad23011b81c8c242b8c52: Status 404 returned error can't find the container with id 06e086f23bcd8914b97606fcca05875573e5433d2dbad23011b81c8c242b8c52 Oct 03 15:09:38 crc kubenswrapper[4959]: I1003 15:09:38.813390 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:09:39 crc kubenswrapper[4959]: I1003 15:09:39.203562 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 15:09:39 crc kubenswrapper[4959]: W1003 15:09:39.206949 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4d9369a_f74d_4c94_8578_41c4cd5beb0d.slice/crio-3f26cd96fc4facac37895b364ce2c8e506d364a2f231a4599cac6ecd90105fba WatchSource:0}: Error finding container 3f26cd96fc4facac37895b364ce2c8e506d364a2f231a4599cac6ecd90105fba: Status 404 returned error can't find the container with id 3f26cd96fc4facac37895b364ce2c8e506d364a2f231a4599cac6ecd90105fba Oct 03 15:09:39 crc kubenswrapper[4959]: I1003 15:09:39.430373 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"363bf001-9d71-4656-ab22-9bc29fbb88c3","Type":"ContainerStarted","Data":"06e086f23bcd8914b97606fcca05875573e5433d2dbad23011b81c8c242b8c52"} Oct 03 15:09:39 crc kubenswrapper[4959]: I1003 15:09:39.433805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c4d9369a-f74d-4c94-8578-41c4cd5beb0d","Type":"ContainerStarted","Data":"3f26cd96fc4facac37895b364ce2c8e506d364a2f231a4599cac6ecd90105fba"} Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.450504 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c4d9369a-f74d-4c94-8578-41c4cd5beb0d","Type":"ContainerStarted","Data":"3ed51ed84166dc10a66223971bf4757aefed76b551d98e49085713c14dfd3580"} Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.455409 4959 generic.go:334] "Generic (PLEG): container finished" podID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerID="b09c3e8db4ebf7cd97490d26e218b883286a7521084748183459aacef13b5c0b" exitCode=0 Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.455493 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerDied","Data":"b09c3e8db4ebf7cd97490d26e218b883286a7521084748183459aacef13b5c0b"} Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.460937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"363bf001-9d71-4656-ab22-9bc29fbb88c3","Type":"ContainerStarted","Data":"2a7346ab8d2d88b1f028b789ed822c9f1f330ac127c259990ca9da9b023d00ef"} Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.460982 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"363bf001-9d71-4656-ab22-9bc29fbb88c3","Type":"ContainerStarted","Data":"cc5bd5f028f9fa414d0ff5c41b086218cb0572db277b20e31e3d050f62bb455c"} Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.491172 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.70746126 podStartE2EDuration="3.491148088s" podCreationTimestamp="2025-10-03 15:09:37 +0000 UTC" firstStartedPulling="2025-10-03 15:09:38.813137494 +0000 UTC m=+5948.016480911" lastFinishedPulling="2025-10-03 15:09:39.596824322 +0000 UTC m=+5948.800167739" observedRunningTime="2025-10-03 15:09:40.486623447 +0000 UTC m=+5949.689966864" watchObservedRunningTime="2025-10-03 15:09:40.491148088 +0000 UTC m=+5949.694491505" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.785934 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.955434 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956150 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956327 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956454 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956511 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956541 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gpgw\" (UniqueName: \"kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.956580 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts\") pod \"fc127260-f1e9-448f-aa89-8f141fc0d927\" (UID: \"fc127260-f1e9-448f-aa89-8f141fc0d927\") " Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.959929 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.960341 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs" (OuterVolumeSpecName: "logs") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.974572 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.980936 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw" (OuterVolumeSpecName: "kube-api-access-7gpgw") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "kube-api-access-7gpgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.988361 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:40 crc kubenswrapper[4959]: I1003 15:09:40.993889 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts" (OuterVolumeSpecName: "scripts") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.024501 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data" (OuterVolumeSpecName: "config-data") pod "fc127260-f1e9-448f-aa89-8f141fc0d927" (UID: "fc127260-f1e9-448f-aa89-8f141fc0d927"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058912 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc127260-f1e9-448f-aa89-8f141fc0d927-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058942 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058954 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058962 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc127260-f1e9-448f-aa89-8f141fc0d927-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058971 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058979 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gpgw\" (UniqueName: \"kubernetes.io/projected/fc127260-f1e9-448f-aa89-8f141fc0d927-kube-api-access-7gpgw\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.058987 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc127260-f1e9-448f-aa89-8f141fc0d927-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.420170 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.474063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c4d9369a-f74d-4c94-8578-41c4cd5beb0d","Type":"ContainerStarted","Data":"a2563585db02a36fdaed8e0afb5391db040fd46fc3791ad02082c171c40356d0"} Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.478162 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.482146 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc127260-f1e9-448f-aa89-8f141fc0d927","Type":"ContainerDied","Data":"07e38c7045a96f0d5f38229f3ea4e2edce318a7e8919cc6997120d2b9de2533d"} Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.482256 4959 scope.go:117] "RemoveContainer" containerID="b09c3e8db4ebf7cd97490d26e218b883286a7521084748183459aacef13b5c0b" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.528846 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.711231624 podStartE2EDuration="3.528825452s" podCreationTimestamp="2025-10-03 15:09:38 +0000 UTC" firstStartedPulling="2025-10-03 15:09:39.208803697 +0000 UTC m=+5948.412147114" lastFinishedPulling="2025-10-03 15:09:40.026397525 +0000 UTC m=+5949.229740942" observedRunningTime="2025-10-03 15:09:41.520463038 +0000 UTC m=+5950.723806505" watchObservedRunningTime="2025-10-03 15:09:41.528825452 +0000 UTC m=+5950.732168879" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.528921 4959 scope.go:117] "RemoveContainer" containerID="96bb4a46a04836ffa6cf9033115d39f2f091bc8e906cb8a8b40d5997ad7a8093" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.560318 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.584686 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.599052 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:41 crc kubenswrapper[4959]: E1003 15:09:41.600978 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api-log" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.601012 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api-log" Oct 03 15:09:41 crc kubenswrapper[4959]: E1003 15:09:41.601067 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.601076 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.601323 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.601355 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api-log" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.602599 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.606346 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.609005 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.708819 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" path="/var/lib/kubelet/pods/fc127260-f1e9-448f-aa89-8f141fc0d927/volumes" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780677 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-logs\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780766 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780785 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-scripts\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780813 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5thx4\" (UniqueName: \"kubernetes.io/projected/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-kube-api-access-5thx4\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780864 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.780900 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.809696 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.811818 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.817453 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.838965 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.839301 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.842283 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.849109 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883133 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883228 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883254 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-logs\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883316 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883338 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-scripts\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5thx4\" (UniqueName: \"kubernetes.io/projected/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-kube-api-access-5thx4\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.883427 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.885267 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-logs\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.889711 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.901791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-scripts\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.901957 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.902406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.902553 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-config-data\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.910420 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5thx4\" (UniqueName: \"kubernetes.io/projected/b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae-kube-api-access-5thx4\") pod \"cinder-api-0\" (UID: \"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae\") " pod="openstack/cinder-api-0" Oct 03 15:09:41 crc kubenswrapper[4959]: I1003 15:09:41.937289 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 15:09:42 crc kubenswrapper[4959]: I1003 15:09:42.443304 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 15:09:42 crc kubenswrapper[4959]: W1003 15:09:42.450345 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5e1778c_6d55_4e02_9b2e_d7ce9f9ebeae.slice/crio-f83f0a3186af5d03584ef1334490de26d4a33f208e9d969caa872fea55b2559a WatchSource:0}: Error finding container f83f0a3186af5d03584ef1334490de26d4a33f208e9d969caa872fea55b2559a: Status 404 returned error can't find the container with id f83f0a3186af5d03584ef1334490de26d4a33f208e9d969caa872fea55b2559a Oct 03 15:09:42 crc kubenswrapper[4959]: I1003 15:09:42.504100 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae","Type":"ContainerStarted","Data":"f83f0a3186af5d03584ef1334490de26d4a33f208e9d969caa872fea55b2559a"} Oct 03 15:09:42 crc kubenswrapper[4959]: I1003 15:09:42.506229 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 15:09:42 crc kubenswrapper[4959]: I1003 15:09:42.511370 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 15:09:42 crc kubenswrapper[4959]: I1003 15:09:42.513296 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 15:09:43 crc kubenswrapper[4959]: I1003 15:09:43.140361 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:43 crc kubenswrapper[4959]: I1003 15:09:43.520335 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae","Type":"ContainerStarted","Data":"12216984e339cc7e047f11eac1f48ac6ad762a83ff9cae807b01d5674ad02913"} Oct 03 15:09:43 crc kubenswrapper[4959]: I1003 15:09:43.607380 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 03 15:09:44 crc kubenswrapper[4959]: I1003 15:09:44.537449 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae","Type":"ContainerStarted","Data":"30a3bfde2597b67c320699f4d756c6b8c291d6e549cf169e77e92d9658f13c53"} Oct 03 15:09:44 crc kubenswrapper[4959]: I1003 15:09:44.538589 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 15:09:44 crc kubenswrapper[4959]: I1003 15:09:44.575650 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.575602813 podStartE2EDuration="3.575602813s" podCreationTimestamp="2025-10-03 15:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:44.56077664 +0000 UTC m=+5953.764120057" watchObservedRunningTime="2025-10-03 15:09:44.575602813 +0000 UTC m=+5953.778946240" Oct 03 15:09:45 crc kubenswrapper[4959]: I1003 15:09:45.683034 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="fc127260-f1e9-448f-aa89-8f141fc0d927" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.83:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 15:09:45 crc kubenswrapper[4959]: I1003 15:09:45.685187 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:09:45 crc kubenswrapper[4959]: E1003 15:09:45.685513 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:09:46 crc kubenswrapper[4959]: I1003 15:09:46.661163 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 15:09:46 crc kubenswrapper[4959]: I1003 15:09:46.717151 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:47 crc kubenswrapper[4959]: I1003 15:09:47.567746 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="cinder-scheduler" containerID="cri-o://9f0c39d9b0b43ca3320034aee406c57377c48e9e141639250c6d80422cf7c680" gracePeriod=30 Oct 03 15:09:47 crc kubenswrapper[4959]: I1003 15:09:47.567844 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="probe" containerID="cri-o://08326680f8abe2848e00a23d38ff42e4293f18ef8c72c26e02af8504afcd4a0a" gracePeriod=30 Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.357102 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.579700 4959 generic.go:334] "Generic (PLEG): container finished" podID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerID="08326680f8abe2848e00a23d38ff42e4293f18ef8c72c26e02af8504afcd4a0a" exitCode=0 Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.580134 4959 generic.go:334] "Generic (PLEG): container finished" podID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerID="9f0c39d9b0b43ca3320034aee406c57377c48e9e141639250c6d80422cf7c680" exitCode=0 Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.580153 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerDied","Data":"08326680f8abe2848e00a23d38ff42e4293f18ef8c72c26e02af8504afcd4a0a"} Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.580180 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerDied","Data":"9f0c39d9b0b43ca3320034aee406c57377c48e9e141639250c6d80422cf7c680"} Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.814359 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.888618 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938052 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgc8d\" (UniqueName: \"kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938243 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938315 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938353 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938380 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id\") pod \"db54e99b-1423-4819-ab9e-9fcc116a7337\" (UID: \"db54e99b-1423-4819-ab9e-9fcc116a7337\") " Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.938914 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.943529 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts" (OuterVolumeSpecName: "scripts") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.943678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d" (OuterVolumeSpecName: "kube-api-access-bgc8d") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "kube-api-access-bgc8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:09:48 crc kubenswrapper[4959]: I1003 15:09:48.943816 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.003408 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.040763 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.040802 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.040811 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.040820 4959 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/db54e99b-1423-4819-ab9e-9fcc116a7337-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.040829 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgc8d\" (UniqueName: \"kubernetes.io/projected/db54e99b-1423-4819-ab9e-9fcc116a7337-kube-api-access-bgc8d\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.052781 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data" (OuterVolumeSpecName: "config-data") pod "db54e99b-1423-4819-ab9e-9fcc116a7337" (UID: "db54e99b-1423-4819-ab9e-9fcc116a7337"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.143314 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db54e99b-1423-4819-ab9e-9fcc116a7337-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.596277 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"db54e99b-1423-4819-ab9e-9fcc116a7337","Type":"ContainerDied","Data":"0a8b468ae333eb960ef77d53290ab72ee0fe48b5386becb2f828cecc0816a80a"} Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.596401 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.596781 4959 scope.go:117] "RemoveContainer" containerID="08326680f8abe2848e00a23d38ff42e4293f18ef8c72c26e02af8504afcd4a0a" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.627968 4959 scope.go:117] "RemoveContainer" containerID="9f0c39d9b0b43ca3320034aee406c57377c48e9e141639250c6d80422cf7c680" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.643017 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.650775 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.665127 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:49 crc kubenswrapper[4959]: E1003 15:09:49.665550 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="cinder-scheduler" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.665568 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="cinder-scheduler" Oct 03 15:09:49 crc kubenswrapper[4959]: E1003 15:09:49.665594 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="probe" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.665601 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="probe" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.665765 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="probe" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.665790 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" containerName="cinder-scheduler" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.666698 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.669602 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.679609 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.698998 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db54e99b-1423-4819-ab9e-9fcc116a7337" path="/var/lib/kubelet/pods/db54e99b-1423-4819-ab9e-9fcc116a7337/volumes" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.751911 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-scripts\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.752272 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.752358 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355c44f-1794-4bb7-9c2d-f71963345c81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.752644 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndrtd\" (UniqueName: \"kubernetes.io/projected/6355c44f-1794-4bb7-9c2d-f71963345c81-kube-api-access-ndrtd\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.752755 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.752984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854598 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854699 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-scripts\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854748 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854775 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355c44f-1794-4bb7-9c2d-f71963345c81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854853 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndrtd\" (UniqueName: \"kubernetes.io/projected/6355c44f-1794-4bb7-9c2d-f71963345c81-kube-api-access-ndrtd\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.854912 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.855160 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355c44f-1794-4bb7-9c2d-f71963345c81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.859762 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.868915 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.868925 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-scripts\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.869612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355c44f-1794-4bb7-9c2d-f71963345c81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.878403 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndrtd\" (UniqueName: \"kubernetes.io/projected/6355c44f-1794-4bb7-9c2d-f71963345c81-kube-api-access-ndrtd\") pod \"cinder-scheduler-0\" (UID: \"6355c44f-1794-4bb7-9c2d-f71963345c81\") " pod="openstack/cinder-scheduler-0" Oct 03 15:09:49 crc kubenswrapper[4959]: I1003 15:09:49.989118 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 15:09:50 crc kubenswrapper[4959]: I1003 15:09:50.494113 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 15:09:50 crc kubenswrapper[4959]: I1003 15:09:50.619417 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6355c44f-1794-4bb7-9c2d-f71963345c81","Type":"ContainerStarted","Data":"11216b5eaf5373535e3ab58f3c107cc05d760fedcd5de3805eaff633e98f94c2"} Oct 03 15:09:51 crc kubenswrapper[4959]: I1003 15:09:51.641220 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6355c44f-1794-4bb7-9c2d-f71963345c81","Type":"ContainerStarted","Data":"1ed34cd6e12b9b2111412136040f27e9d9519e0b0d13a21e3090d74712d7b175"} Oct 03 15:09:51 crc kubenswrapper[4959]: I1003 15:09:51.674467 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.674435036 podStartE2EDuration="2.674435036s" podCreationTimestamp="2025-10-03 15:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:09:51.658495546 +0000 UTC m=+5960.861839013" watchObservedRunningTime="2025-10-03 15:09:51.674435036 +0000 UTC m=+5960.877778483" Oct 03 15:09:52 crc kubenswrapper[4959]: I1003 15:09:52.668041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6355c44f-1794-4bb7-9c2d-f71963345c81","Type":"ContainerStarted","Data":"a514743e9b2aaccc4a0c1cf9cf15b18c833ea8ed06499ad0c01cf4c970dd4c79"} Oct 03 15:09:53 crc kubenswrapper[4959]: I1003 15:09:53.846957 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 03 15:09:54 crc kubenswrapper[4959]: I1003 15:09:54.989543 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 15:09:56 crc kubenswrapper[4959]: I1003 15:09:56.686066 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:09:56 crc kubenswrapper[4959]: E1003 15:09:56.686970 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:10:00 crc kubenswrapper[4959]: I1003 15:10:00.239214 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 15:10:08 crc kubenswrapper[4959]: I1003 15:10:08.685542 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:10:09 crc kubenswrapper[4959]: I1003 15:10:09.840620 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a"} Oct 03 15:10:21 crc kubenswrapper[4959]: I1003 15:10:21.074309 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ql2nc"] Oct 03 15:10:21 crc kubenswrapper[4959]: I1003 15:10:21.086810 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ql2nc"] Oct 03 15:10:21 crc kubenswrapper[4959]: I1003 15:10:21.706057 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f54f365-2baf-43d8-b046-c316d136ea1d" path="/var/lib/kubelet/pods/3f54f365-2baf-43d8-b046-c316d136ea1d/volumes" Oct 03 15:10:31 crc kubenswrapper[4959]: I1003 15:10:31.027707 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f61d-account-create-fwfrw"] Oct 03 15:10:31 crc kubenswrapper[4959]: I1003 15:10:31.038901 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f61d-account-create-fwfrw"] Oct 03 15:10:31 crc kubenswrapper[4959]: I1003 15:10:31.701059 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1918b5f3-9683-44e7-ad21-f7cb2935cd7d" path="/var/lib/kubelet/pods/1918b5f3-9683-44e7-ad21-f7cb2935cd7d/volumes" Oct 03 15:10:38 crc kubenswrapper[4959]: I1003 15:10:38.037968 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jbwnh"] Oct 03 15:10:38 crc kubenswrapper[4959]: I1003 15:10:38.055507 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jbwnh"] Oct 03 15:10:39 crc kubenswrapper[4959]: I1003 15:10:39.695181 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a234cd-5df5-4e16-88fe-78221ec12175" path="/var/lib/kubelet/pods/54a234cd-5df5-4e16-88fe-78221ec12175/volumes" Oct 03 15:10:52 crc kubenswrapper[4959]: I1003 15:10:52.048784 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5vgl4"] Oct 03 15:10:52 crc kubenswrapper[4959]: I1003 15:10:52.058688 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5vgl4"] Oct 03 15:10:53 crc kubenswrapper[4959]: I1003 15:10:53.712566 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c398778d-bdd5-49c1-a1b8-348cd53cf86a" path="/var/lib/kubelet/pods/c398778d-bdd5-49c1-a1b8-348cd53cf86a/volumes" Oct 03 15:11:06 crc kubenswrapper[4959]: I1003 15:11:06.485557 4959 scope.go:117] "RemoveContainer" containerID="134d4662d6c6e20af268d24f25cb84fcd5ad640260002e4a4fc2c65c92e99e85" Oct 03 15:11:06 crc kubenswrapper[4959]: I1003 15:11:06.532012 4959 scope.go:117] "RemoveContainer" containerID="0955324328fc494d1dd9366aaae06635e38e3be945ae9642faef62cf6ead7456" Oct 03 15:11:06 crc kubenswrapper[4959]: I1003 15:11:06.566215 4959 scope.go:117] "RemoveContainer" containerID="fc37da1a06d7eeb6bcef1763a3c1a1d2f2b9b23367e175f57fa2b64d5b4f8bf7" Oct 03 15:11:06 crc kubenswrapper[4959]: I1003 15:11:06.593923 4959 scope.go:117] "RemoveContainer" containerID="471603a8887b5603e091aeec9862021dc392533e8776b701f1b661441c62771a" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.698762 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8nphv"] Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.700946 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.703449 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-kdrkl" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.704527 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.715597 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-p6zgv"] Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.718040 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.727890 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8nphv"] Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.737956 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p6zgv"] Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.802625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-log-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.802704 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqtpc\" (UniqueName: \"kubernetes.io/projected/973e0edf-ad3a-4987-a8e6-4724186d830d-kube-api-access-dqtpc\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.802734 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973e0edf-ad3a-4987-a8e6-4724186d830d-scripts\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.802760 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.802780 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904351 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqtpc\" (UniqueName: \"kubernetes.io/projected/973e0edf-ad3a-4987-a8e6-4724186d830d-kube-api-access-dqtpc\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904437 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973e0edf-ad3a-4987-a8e6-4724186d830d-scripts\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904498 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrw2h\" (UniqueName: \"kubernetes.io/projected/9b1dcf14-6652-4948-a200-d533be359fd0-kube-api-access-lrw2h\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904581 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-run\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904613 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904724 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-lib\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904775 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-etc-ovs\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904826 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b1dcf14-6652-4948-a200-d533be359fd0-scripts\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.904924 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-log-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.905018 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-log\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.905532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.905652 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-log-ovn\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.905740 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/973e0edf-ad3a-4987-a8e6-4724186d830d-var-run\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.907663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973e0edf-ad3a-4987-a8e6-4724186d830d-scripts\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:40 crc kubenswrapper[4959]: I1003 15:11:40.933183 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqtpc\" (UniqueName: \"kubernetes.io/projected/973e0edf-ad3a-4987-a8e6-4724186d830d-kube-api-access-dqtpc\") pod \"ovn-controller-8nphv\" (UID: \"973e0edf-ad3a-4987-a8e6-4724186d830d\") " pod="openstack/ovn-controller-8nphv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006398 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-log\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006463 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrw2h\" (UniqueName: \"kubernetes.io/projected/9b1dcf14-6652-4948-a200-d533be359fd0-kube-api-access-lrw2h\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006493 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-run\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006542 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-lib\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006575 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-etc-ovs\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006602 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b1dcf14-6652-4948-a200-d533be359fd0-scripts\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006688 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-log\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.006756 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-run\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.007063 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-var-lib\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.007232 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9b1dcf14-6652-4948-a200-d533be359fd0-etc-ovs\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.009112 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b1dcf14-6652-4948-a200-d533be359fd0-scripts\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.034083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrw2h\" (UniqueName: \"kubernetes.io/projected/9b1dcf14-6652-4948-a200-d533be359fd0-kube-api-access-lrw2h\") pod \"ovn-controller-ovs-p6zgv\" (UID: \"9b1dcf14-6652-4948-a200-d533be359fd0\") " pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.038000 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.050415 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.494363 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8nphv"] Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.824839 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8nphv" event={"ID":"973e0edf-ad3a-4987-a8e6-4724186d830d","Type":"ContainerStarted","Data":"2641e68d6de76756fc8aed6d0ffab91101bc8fdf49f348ee1a76644527a32890"} Oct 03 15:11:41 crc kubenswrapper[4959]: I1003 15:11:41.908176 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-p6zgv"] Oct 03 15:11:41 crc kubenswrapper[4959]: W1003 15:11:41.908533 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b1dcf14_6652_4948_a200_d533be359fd0.slice/crio-8f15c0bf5e4e2a4edcf83f7b42ae9e7e5e8ccc21d396b26432cfbf6b524a98a5 WatchSource:0}: Error finding container 8f15c0bf5e4e2a4edcf83f7b42ae9e7e5e8ccc21d396b26432cfbf6b524a98a5: Status 404 returned error can't find the container with id 8f15c0bf5e4e2a4edcf83f7b42ae9e7e5e8ccc21d396b26432cfbf6b524a98a5 Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.329469 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-zpqwq"] Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.332410 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.340587 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.351162 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zpqwq"] Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.434754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6269efac-eb85-49fe-b427-4b5cc0d50fdd-config\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.434822 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmcms\" (UniqueName: \"kubernetes.io/projected/6269efac-eb85-49fe-b427-4b5cc0d50fdd-kube-api-access-qmcms\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.434868 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovn-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.435055 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovs-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.536698 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6269efac-eb85-49fe-b427-4b5cc0d50fdd-config\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.536771 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmcms\" (UniqueName: \"kubernetes.io/projected/6269efac-eb85-49fe-b427-4b5cc0d50fdd-kube-api-access-qmcms\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.536818 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovn-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.536864 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovs-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.537214 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovs-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.537214 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6269efac-eb85-49fe-b427-4b5cc0d50fdd-ovn-rundir\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.537411 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6269efac-eb85-49fe-b427-4b5cc0d50fdd-config\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.557168 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmcms\" (UniqueName: \"kubernetes.io/projected/6269efac-eb85-49fe-b427-4b5cc0d50fdd-kube-api-access-qmcms\") pod \"ovn-controller-metrics-zpqwq\" (UID: \"6269efac-eb85-49fe-b427-4b5cc0d50fdd\") " pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.683742 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zpqwq" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.839404 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8nphv" event={"ID":"973e0edf-ad3a-4987-a8e6-4724186d830d","Type":"ContainerStarted","Data":"748a284410b17d0fed302145d8791e9aba48d55517187288745170d069e53980"} Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.839865 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8nphv" Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.842027 4959 generic.go:334] "Generic (PLEG): container finished" podID="9b1dcf14-6652-4948-a200-d533be359fd0" containerID="672939e325dfc0305366ace2ff9acd86cd282a6d722f4af38074f0e3a4cd32c4" exitCode=0 Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.842098 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p6zgv" event={"ID":"9b1dcf14-6652-4948-a200-d533be359fd0","Type":"ContainerDied","Data":"672939e325dfc0305366ace2ff9acd86cd282a6d722f4af38074f0e3a4cd32c4"} Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.842142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p6zgv" event={"ID":"9b1dcf14-6652-4948-a200-d533be359fd0","Type":"ContainerStarted","Data":"8f15c0bf5e4e2a4edcf83f7b42ae9e7e5e8ccc21d396b26432cfbf6b524a98a5"} Oct 03 15:11:42 crc kubenswrapper[4959]: I1003 15:11:42.858607 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8nphv" podStartSLOduration=2.858588189 podStartE2EDuration="2.858588189s" podCreationTimestamp="2025-10-03 15:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:11:42.856721233 +0000 UTC m=+6072.060064660" watchObservedRunningTime="2025-10-03 15:11:42.858588189 +0000 UTC m=+6072.061931606" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.184566 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-js7ld"] Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.186076 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.193592 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-js7ld"] Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.269390 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zpqwq"] Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.350611 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87rmn\" (UniqueName: \"kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn\") pod \"octavia-db-create-js7ld\" (UID: \"c285ce72-79d4-44b8-aa1a-f45d6b595bab\") " pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.452311 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87rmn\" (UniqueName: \"kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn\") pod \"octavia-db-create-js7ld\" (UID: \"c285ce72-79d4-44b8-aa1a-f45d6b595bab\") " pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.471464 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87rmn\" (UniqueName: \"kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn\") pod \"octavia-db-create-js7ld\" (UID: \"c285ce72-79d4-44b8-aa1a-f45d6b595bab\") " pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.538215 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.860597 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zpqwq" event={"ID":"6269efac-eb85-49fe-b427-4b5cc0d50fdd","Type":"ContainerStarted","Data":"a34be2122fcf1da9ac319309fa220661ca16289f8e02c7bdbb7829c90c7ad1ce"} Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.861063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zpqwq" event={"ID":"6269efac-eb85-49fe-b427-4b5cc0d50fdd","Type":"ContainerStarted","Data":"95e4894976f96d7fee589cb27a33984f70311551c8da9aad3656843cbb8e2ca6"} Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.864063 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p6zgv" event={"ID":"9b1dcf14-6652-4948-a200-d533be359fd0","Type":"ContainerStarted","Data":"160c2030bbbf1254034695b73d0ab3004ff6523d5c5ea86d132146d136e4f5e5"} Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.864088 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-p6zgv" event={"ID":"9b1dcf14-6652-4948-a200-d533be359fd0","Type":"ContainerStarted","Data":"68b2d25e80ca9589272a76572fa34fb8e04d2a86413a74f409753605d0c22a95"} Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.864235 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.882788 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-zpqwq" podStartSLOduration=1.882766103 podStartE2EDuration="1.882766103s" podCreationTimestamp="2025-10-03 15:11:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:11:43.879343209 +0000 UTC m=+6073.082686626" watchObservedRunningTime="2025-10-03 15:11:43.882766103 +0000 UTC m=+6073.086109520" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.910426 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-p6zgv" podStartSLOduration=3.910399928 podStartE2EDuration="3.910399928s" podCreationTimestamp="2025-10-03 15:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:11:43.909664271 +0000 UTC m=+6073.113007708" watchObservedRunningTime="2025-10-03 15:11:43.910399928 +0000 UTC m=+6073.113743375" Oct 03 15:11:43 crc kubenswrapper[4959]: I1003 15:11:43.975845 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-js7ld"] Oct 03 15:11:44 crc kubenswrapper[4959]: I1003 15:11:44.878894 4959 generic.go:334] "Generic (PLEG): container finished" podID="c285ce72-79d4-44b8-aa1a-f45d6b595bab" containerID="91773be05ca8ae836e57c8108eebfd3b9616b4a689b4db37cdbe8fb9e63a9fea" exitCode=0 Oct 03 15:11:44 crc kubenswrapper[4959]: I1003 15:11:44.878959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-js7ld" event={"ID":"c285ce72-79d4-44b8-aa1a-f45d6b595bab","Type":"ContainerDied","Data":"91773be05ca8ae836e57c8108eebfd3b9616b4a689b4db37cdbe8fb9e63a9fea"} Oct 03 15:11:44 crc kubenswrapper[4959]: I1003 15:11:44.879325 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-js7ld" event={"ID":"c285ce72-79d4-44b8-aa1a-f45d6b595bab","Type":"ContainerStarted","Data":"cd36f02080992d7b1f378c822b553e636ae567a12f2a8c011b94b7883144b53c"} Oct 03 15:11:44 crc kubenswrapper[4959]: I1003 15:11:44.879967 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.305480 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.410736 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87rmn\" (UniqueName: \"kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn\") pod \"c285ce72-79d4-44b8-aa1a-f45d6b595bab\" (UID: \"c285ce72-79d4-44b8-aa1a-f45d6b595bab\") " Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.416971 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn" (OuterVolumeSpecName: "kube-api-access-87rmn") pod "c285ce72-79d4-44b8-aa1a-f45d6b595bab" (UID: "c285ce72-79d4-44b8-aa1a-f45d6b595bab"). InnerVolumeSpecName "kube-api-access-87rmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.514153 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87rmn\" (UniqueName: \"kubernetes.io/projected/c285ce72-79d4-44b8-aa1a-f45d6b595bab-kube-api-access-87rmn\") on node \"crc\" DevicePath \"\"" Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.906233 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-js7ld" event={"ID":"c285ce72-79d4-44b8-aa1a-f45d6b595bab","Type":"ContainerDied","Data":"cd36f02080992d7b1f378c822b553e636ae567a12f2a8c011b94b7883144b53c"} Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.906288 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-js7ld" Oct 03 15:11:46 crc kubenswrapper[4959]: I1003 15:11:46.906347 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd36f02080992d7b1f378c822b553e636ae567a12f2a8c011b94b7883144b53c" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.404532 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-e9e2-account-create-dkfjk"] Oct 03 15:11:55 crc kubenswrapper[4959]: E1003 15:11:55.405674 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c285ce72-79d4-44b8-aa1a-f45d6b595bab" containerName="mariadb-database-create" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.405700 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c285ce72-79d4-44b8-aa1a-f45d6b595bab" containerName="mariadb-database-create" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.405985 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c285ce72-79d4-44b8-aa1a-f45d6b595bab" containerName="mariadb-database-create" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.406885 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.410390 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.427373 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e9e2-account-create-dkfjk"] Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.485165 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc7zm\" (UniqueName: \"kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm\") pod \"octavia-e9e2-account-create-dkfjk\" (UID: \"de58b2a6-3503-4b49-bf4a-c393087c4f47\") " pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.587920 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc7zm\" (UniqueName: \"kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm\") pod \"octavia-e9e2-account-create-dkfjk\" (UID: \"de58b2a6-3503-4b49-bf4a-c393087c4f47\") " pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.607089 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc7zm\" (UniqueName: \"kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm\") pod \"octavia-e9e2-account-create-dkfjk\" (UID: \"de58b2a6-3503-4b49-bf4a-c393087c4f47\") " pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:55 crc kubenswrapper[4959]: I1003 15:11:55.732410 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:56 crc kubenswrapper[4959]: I1003 15:11:56.184522 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e9e2-account-create-dkfjk"] Oct 03 15:11:56 crc kubenswrapper[4959]: W1003 15:11:56.186460 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde58b2a6_3503_4b49_bf4a_c393087c4f47.slice/crio-e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51 WatchSource:0}: Error finding container e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51: Status 404 returned error can't find the container with id e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51 Oct 03 15:11:57 crc kubenswrapper[4959]: I1003 15:11:57.008034 4959 generic.go:334] "Generic (PLEG): container finished" podID="de58b2a6-3503-4b49-bf4a-c393087c4f47" containerID="eaaa035a20b4fa311f28e051b156b075223070ad9a7064fd55139278380ee8ac" exitCode=0 Oct 03 15:11:57 crc kubenswrapper[4959]: I1003 15:11:57.008100 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e9e2-account-create-dkfjk" event={"ID":"de58b2a6-3503-4b49-bf4a-c393087c4f47","Type":"ContainerDied","Data":"eaaa035a20b4fa311f28e051b156b075223070ad9a7064fd55139278380ee8ac"} Oct 03 15:11:57 crc kubenswrapper[4959]: I1003 15:11:57.008387 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e9e2-account-create-dkfjk" event={"ID":"de58b2a6-3503-4b49-bf4a-c393087c4f47","Type":"ContainerStarted","Data":"e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51"} Oct 03 15:11:58 crc kubenswrapper[4959]: I1003 15:11:58.307351 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:58 crc kubenswrapper[4959]: I1003 15:11:58.439214 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc7zm\" (UniqueName: \"kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm\") pod \"de58b2a6-3503-4b49-bf4a-c393087c4f47\" (UID: \"de58b2a6-3503-4b49-bf4a-c393087c4f47\") " Oct 03 15:11:58 crc kubenswrapper[4959]: I1003 15:11:58.450558 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm" (OuterVolumeSpecName: "kube-api-access-sc7zm") pod "de58b2a6-3503-4b49-bf4a-c393087c4f47" (UID: "de58b2a6-3503-4b49-bf4a-c393087c4f47"). InnerVolumeSpecName "kube-api-access-sc7zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:11:58 crc kubenswrapper[4959]: I1003 15:11:58.541476 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc7zm\" (UniqueName: \"kubernetes.io/projected/de58b2a6-3503-4b49-bf4a-c393087c4f47-kube-api-access-sc7zm\") on node \"crc\" DevicePath \"\"" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.031814 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e9e2-account-create-dkfjk" event={"ID":"de58b2a6-3503-4b49-bf4a-c393087c4f47","Type":"ContainerDied","Data":"e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51"} Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.031864 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e9e2-account-create-dkfjk" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.031875 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20c53036ff2fb7d577a3d39850ea02d04efcdb93ea41fe36a9271e6d26fff51" Oct 03 15:11:59 crc kubenswrapper[4959]: E1003 15:11:59.134723 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde58b2a6_3503_4b49_bf4a_c393087c4f47.slice\": RecentStats: unable to find data in memory cache]" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.764291 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:11:59 crc kubenswrapper[4959]: E1003 15:11:59.765050 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de58b2a6-3503-4b49-bf4a-c393087c4f47" containerName="mariadb-account-create" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.765065 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="de58b2a6-3503-4b49-bf4a-c393087c4f47" containerName="mariadb-account-create" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.765343 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="de58b2a6-3503-4b49-bf4a-c393087c4f47" containerName="mariadb-account-create" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.767025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.805769 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.873660 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.873758 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v72w2\" (UniqueName: \"kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.873823 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.975285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v72w2\" (UniqueName: \"kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.975623 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.975749 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.976129 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.976296 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:11:59 crc kubenswrapper[4959]: I1003 15:11:59.998758 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v72w2\" (UniqueName: \"kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2\") pod \"certified-operators-r6hxw\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:00 crc kubenswrapper[4959]: I1003 15:12:00.098450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:00 crc kubenswrapper[4959]: I1003 15:12:00.630223 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:12:00 crc kubenswrapper[4959]: W1003 15:12:00.631335 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18447ff8_d783_436d_9c2a_8c49d7beb4d8.slice/crio-6866d26128e3e7e9f3893f75f862894dba39a8deda713cfcf07f1e3da3b35c70 WatchSource:0}: Error finding container 6866d26128e3e7e9f3893f75f862894dba39a8deda713cfcf07f1e3da3b35c70: Status 404 returned error can't find the container with id 6866d26128e3e7e9f3893f75f862894dba39a8deda713cfcf07f1e3da3b35c70 Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.051831 4959 generic.go:334] "Generic (PLEG): container finished" podID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerID="c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913" exitCode=0 Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.051909 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerDied","Data":"c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913"} Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.052164 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerStarted","Data":"6866d26128e3e7e9f3893f75f862894dba39a8deda713cfcf07f1e3da3b35c70"} Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.438695 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-tvbsr"] Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.441308 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.447765 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-tvbsr"] Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.505322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czt8\" (UniqueName: \"kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8\") pod \"octavia-persistence-db-create-tvbsr\" (UID: \"670af6f4-13ae-406a-b0db-8a84ada6ea30\") " pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.608377 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czt8\" (UniqueName: \"kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8\") pod \"octavia-persistence-db-create-tvbsr\" (UID: \"670af6f4-13ae-406a-b0db-8a84ada6ea30\") " pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.628182 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czt8\" (UniqueName: \"kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8\") pod \"octavia-persistence-db-create-tvbsr\" (UID: \"670af6f4-13ae-406a-b0db-8a84ada6ea30\") " pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:01 crc kubenswrapper[4959]: I1003 15:12:01.769062 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:02 crc kubenswrapper[4959]: I1003 15:12:02.214397 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-tvbsr"] Oct 03 15:12:02 crc kubenswrapper[4959]: W1003 15:12:02.218499 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod670af6f4_13ae_406a_b0db_8a84ada6ea30.slice/crio-dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390 WatchSource:0}: Error finding container dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390: Status 404 returned error can't find the container with id dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390 Oct 03 15:12:03 crc kubenswrapper[4959]: I1003 15:12:03.073184 4959 generic.go:334] "Generic (PLEG): container finished" podID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerID="e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52" exitCode=0 Oct 03 15:12:03 crc kubenswrapper[4959]: I1003 15:12:03.073268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerDied","Data":"e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52"} Oct 03 15:12:03 crc kubenswrapper[4959]: I1003 15:12:03.076043 4959 generic.go:334] "Generic (PLEG): container finished" podID="670af6f4-13ae-406a-b0db-8a84ada6ea30" containerID="4b0786610bbe66998aa9ab2a84ef623f2abf04f98f9ef3aa67a70e6363d09933" exitCode=0 Oct 03 15:12:03 crc kubenswrapper[4959]: I1003 15:12:03.076085 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-tvbsr" event={"ID":"670af6f4-13ae-406a-b0db-8a84ada6ea30","Type":"ContainerDied","Data":"4b0786610bbe66998aa9ab2a84ef623f2abf04f98f9ef3aa67a70e6363d09933"} Oct 03 15:12:03 crc kubenswrapper[4959]: I1003 15:12:03.076120 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-tvbsr" event={"ID":"670af6f4-13ae-406a-b0db-8a84ada6ea30","Type":"ContainerStarted","Data":"dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390"} Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.086304 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerStarted","Data":"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b"} Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.114395 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r6hxw" podStartSLOduration=2.450792646 podStartE2EDuration="5.11436672s" podCreationTimestamp="2025-10-03 15:11:59 +0000 UTC" firstStartedPulling="2025-10-03 15:12:01.053611437 +0000 UTC m=+6090.256954854" lastFinishedPulling="2025-10-03 15:12:03.717185511 +0000 UTC m=+6092.920528928" observedRunningTime="2025-10-03 15:12:04.101962236 +0000 UTC m=+6093.305305663" watchObservedRunningTime="2025-10-03 15:12:04.11436672 +0000 UTC m=+6093.317710127" Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.438957 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.583142 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8czt8\" (UniqueName: \"kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8\") pod \"670af6f4-13ae-406a-b0db-8a84ada6ea30\" (UID: \"670af6f4-13ae-406a-b0db-8a84ada6ea30\") " Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.591997 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8" (OuterVolumeSpecName: "kube-api-access-8czt8") pod "670af6f4-13ae-406a-b0db-8a84ada6ea30" (UID: "670af6f4-13ae-406a-b0db-8a84ada6ea30"). InnerVolumeSpecName "kube-api-access-8czt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:12:04 crc kubenswrapper[4959]: I1003 15:12:04.685152 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8czt8\" (UniqueName: \"kubernetes.io/projected/670af6f4-13ae-406a-b0db-8a84ada6ea30-kube-api-access-8czt8\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:05 crc kubenswrapper[4959]: I1003 15:12:05.097124 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-tvbsr" event={"ID":"670af6f4-13ae-406a-b0db-8a84ada6ea30","Type":"ContainerDied","Data":"dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390"} Oct 03 15:12:05 crc kubenswrapper[4959]: I1003 15:12:05.097501 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd0fffe9b92b1fc49f23b012fda23a2c68506a824e7d1c7e54fe1c6d5d9fe390" Oct 03 15:12:05 crc kubenswrapper[4959]: I1003 15:12:05.097145 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-tvbsr" Oct 03 15:12:10 crc kubenswrapper[4959]: I1003 15:12:10.099071 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:10 crc kubenswrapper[4959]: I1003 15:12:10.099704 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:10 crc kubenswrapper[4959]: I1003 15:12:10.171352 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:10 crc kubenswrapper[4959]: I1003 15:12:10.230582 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:10 crc kubenswrapper[4959]: I1003 15:12:10.416732 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:12:11 crc kubenswrapper[4959]: I1003 15:12:11.111630 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8nphv" podUID="973e0edf-ad3a-4987-a8e6-4724186d830d" containerName="ovn-controller" probeResult="failure" output=< Oct 03 15:12:11 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 15:12:11 crc kubenswrapper[4959]: > Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.171243 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r6hxw" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="registry-server" containerID="cri-o://3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b" gracePeriod=2 Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.512988 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-d0ae-account-create-cf4c4"] Oct 03 15:12:12 crc kubenswrapper[4959]: E1003 15:12:12.513713 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670af6f4-13ae-406a-b0db-8a84ada6ea30" containerName="mariadb-database-create" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.513725 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="670af6f4-13ae-406a-b0db-8a84ada6ea30" containerName="mariadb-database-create" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.513954 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="670af6f4-13ae-406a-b0db-8a84ada6ea30" containerName="mariadb-database-create" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.514633 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.516725 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.524110 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-d0ae-account-create-cf4c4"] Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.541962 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jjrj\" (UniqueName: \"kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj\") pod \"octavia-d0ae-account-create-cf4c4\" (UID: \"f8c90709-b50f-4c10-af62-f646f51fad07\") " pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.643161 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jjrj\" (UniqueName: \"kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj\") pod \"octavia-d0ae-account-create-cf4c4\" (UID: \"f8c90709-b50f-4c10-af62-f646f51fad07\") " pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.663952 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.666178 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jjrj\" (UniqueName: \"kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj\") pod \"octavia-d0ae-account-create-cf4c4\" (UID: \"f8c90709-b50f-4c10-af62-f646f51fad07\") " pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.744934 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities\") pod \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.745665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v72w2\" (UniqueName: \"kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2\") pod \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.745740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities" (OuterVolumeSpecName: "utilities") pod "18447ff8-d783-436d-9c2a-8c49d7beb4d8" (UID: "18447ff8-d783-436d-9c2a-8c49d7beb4d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.745781 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content\") pod \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\" (UID: \"18447ff8-d783-436d-9c2a-8c49d7beb4d8\") " Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.746402 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.749492 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2" (OuterVolumeSpecName: "kube-api-access-v72w2") pod "18447ff8-d783-436d-9c2a-8c49d7beb4d8" (UID: "18447ff8-d783-436d-9c2a-8c49d7beb4d8"). InnerVolumeSpecName "kube-api-access-v72w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.792871 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18447ff8-d783-436d-9c2a-8c49d7beb4d8" (UID: "18447ff8-d783-436d-9c2a-8c49d7beb4d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.837286 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.848045 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v72w2\" (UniqueName: \"kubernetes.io/projected/18447ff8-d783-436d-9c2a-8c49d7beb4d8-kube-api-access-v72w2\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:12 crc kubenswrapper[4959]: I1003 15:12:12.848080 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18447ff8-d783-436d-9c2a-8c49d7beb4d8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.183606 4959 generic.go:334] "Generic (PLEG): container finished" podID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerID="3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b" exitCode=0 Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.183667 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6hxw" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.183676 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerDied","Data":"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b"} Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.184018 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6hxw" event={"ID":"18447ff8-d783-436d-9c2a-8c49d7beb4d8","Type":"ContainerDied","Data":"6866d26128e3e7e9f3893f75f862894dba39a8deda713cfcf07f1e3da3b35c70"} Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.184049 4959 scope.go:117] "RemoveContainer" containerID="3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.219764 4959 scope.go:117] "RemoveContainer" containerID="e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.223352 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.233871 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r6hxw"] Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.247418 4959 scope.go:117] "RemoveContainer" containerID="c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.279170 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-d0ae-account-create-cf4c4"] Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.298263 4959 scope.go:117] "RemoveContainer" containerID="3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b" Oct 03 15:12:13 crc kubenswrapper[4959]: E1003 15:12:13.299249 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b\": container with ID starting with 3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b not found: ID does not exist" containerID="3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.299287 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b"} err="failed to get container status \"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b\": rpc error: code = NotFound desc = could not find container \"3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b\": container with ID starting with 3cf03d2cce9c34f07859a18a7dca34213a35e79eee4e2584fc4e1a0b3cbaf87b not found: ID does not exist" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.299317 4959 scope.go:117] "RemoveContainer" containerID="e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52" Oct 03 15:12:13 crc kubenswrapper[4959]: E1003 15:12:13.300322 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52\": container with ID starting with e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52 not found: ID does not exist" containerID="e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.300358 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52"} err="failed to get container status \"e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52\": rpc error: code = NotFound desc = could not find container \"e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52\": container with ID starting with e19433570cb5030ae25b42fb9dba80b7068f031114396402176f52a3e1a9dc52 not found: ID does not exist" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.300381 4959 scope.go:117] "RemoveContainer" containerID="c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913" Oct 03 15:12:13 crc kubenswrapper[4959]: E1003 15:12:13.300637 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913\": container with ID starting with c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913 not found: ID does not exist" containerID="c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.300662 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913"} err="failed to get container status \"c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913\": rpc error: code = NotFound desc = could not find container \"c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913\": container with ID starting with c8efbe61cbe8d15aea6d375147e3eb12c24d0dfc7fe7579b5b5290fee63a8913 not found: ID does not exist" Oct 03 15:12:13 crc kubenswrapper[4959]: I1003 15:12:13.697699 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" path="/var/lib/kubelet/pods/18447ff8-d783-436d-9c2a-8c49d7beb4d8/volumes" Oct 03 15:12:14 crc kubenswrapper[4959]: I1003 15:12:14.195958 4959 generic.go:334] "Generic (PLEG): container finished" podID="f8c90709-b50f-4c10-af62-f646f51fad07" containerID="316abf3d476478c428b230de8b4a497af167965d7b3f3a6fa83402d95e383d1f" exitCode=0 Oct 03 15:12:14 crc kubenswrapper[4959]: I1003 15:12:14.196013 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-d0ae-account-create-cf4c4" event={"ID":"f8c90709-b50f-4c10-af62-f646f51fad07","Type":"ContainerDied","Data":"316abf3d476478c428b230de8b4a497af167965d7b3f3a6fa83402d95e383d1f"} Oct 03 15:12:14 crc kubenswrapper[4959]: I1003 15:12:14.196080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-d0ae-account-create-cf4c4" event={"ID":"f8c90709-b50f-4c10-af62-f646f51fad07","Type":"ContainerStarted","Data":"4eb72cf7cd45bc7b3b448f126fc1dfaa9c53ae3a492afd0379f176df5b5e3701"} Oct 03 15:12:15 crc kubenswrapper[4959]: I1003 15:12:15.561006 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:15 crc kubenswrapper[4959]: I1003 15:12:15.605937 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jjrj\" (UniqueName: \"kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj\") pod \"f8c90709-b50f-4c10-af62-f646f51fad07\" (UID: \"f8c90709-b50f-4c10-af62-f646f51fad07\") " Oct 03 15:12:15 crc kubenswrapper[4959]: I1003 15:12:15.614588 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj" (OuterVolumeSpecName: "kube-api-access-5jjrj") pod "f8c90709-b50f-4c10-af62-f646f51fad07" (UID: "f8c90709-b50f-4c10-af62-f646f51fad07"). InnerVolumeSpecName "kube-api-access-5jjrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:12:15 crc kubenswrapper[4959]: I1003 15:12:15.711294 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jjrj\" (UniqueName: \"kubernetes.io/projected/f8c90709-b50f-4c10-af62-f646f51fad07-kube-api-access-5jjrj\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.099089 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.113525 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-p6zgv" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.116563 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8nphv" podUID="973e0edf-ad3a-4987-a8e6-4724186d830d" containerName="ovn-controller" probeResult="failure" output=< Oct 03 15:12:16 crc kubenswrapper[4959]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 15:12:16 crc kubenswrapper[4959]: > Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.215223 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-d0ae-account-create-cf4c4" event={"ID":"f8c90709-b50f-4c10-af62-f646f51fad07","Type":"ContainerDied","Data":"4eb72cf7cd45bc7b3b448f126fc1dfaa9c53ae3a492afd0379f176df5b5e3701"} Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.215261 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-d0ae-account-create-cf4c4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.215272 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eb72cf7cd45bc7b3b448f126fc1dfaa9c53ae3a492afd0379f176df5b5e3701" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.233971 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8nphv-config-hdvr4"] Oct 03 15:12:16 crc kubenswrapper[4959]: E1003 15:12:16.234602 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="extract-utilities" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.234680 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="extract-utilities" Oct 03 15:12:16 crc kubenswrapper[4959]: E1003 15:12:16.234776 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="extract-content" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.234833 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="extract-content" Oct 03 15:12:16 crc kubenswrapper[4959]: E1003 15:12:16.234901 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c90709-b50f-4c10-af62-f646f51fad07" containerName="mariadb-account-create" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.234962 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c90709-b50f-4c10-af62-f646f51fad07" containerName="mariadb-account-create" Oct 03 15:12:16 crc kubenswrapper[4959]: E1003 15:12:16.235028 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="registry-server" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.235078 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="registry-server" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.235322 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c90709-b50f-4c10-af62-f646f51fad07" containerName="mariadb-account-create" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.235409 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="18447ff8-d783-436d-9c2a-8c49d7beb4d8" containerName="registry-server" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.236055 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.238037 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.257480 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8nphv-config-hdvr4"] Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320202 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwg44\" (UniqueName: \"kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320468 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320577 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320703 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320839 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.320932 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.422790 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.422877 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwg44\" (UniqueName: \"kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423104 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423185 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423220 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423366 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423404 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.423469 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.424375 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.425399 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.443489 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwg44\" (UniqueName: \"kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44\") pod \"ovn-controller-8nphv-config-hdvr4\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:16 crc kubenswrapper[4959]: I1003 15:12:16.553722 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:17 crc kubenswrapper[4959]: I1003 15:12:17.001023 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8nphv-config-hdvr4"] Oct 03 15:12:17 crc kubenswrapper[4959]: I1003 15:12:17.227748 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8nphv-config-hdvr4" event={"ID":"40cf0a27-46ba-4d51-adb5-5365f40da30d","Type":"ContainerStarted","Data":"bc445a9fdb53d1cbf8840048c29f348da76fd47a511e4110a072b79fa1d2708d"} Oct 03 15:12:18 crc kubenswrapper[4959]: I1003 15:12:18.239623 4959 generic.go:334] "Generic (PLEG): container finished" podID="40cf0a27-46ba-4d51-adb5-5365f40da30d" containerID="a8a8ccb2c084ae66d812cc1ec4033b06a20329b0cf13dbecaf9788172090a69d" exitCode=0 Oct 03 15:12:18 crc kubenswrapper[4959]: I1003 15:12:18.239701 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8nphv-config-hdvr4" event={"ID":"40cf0a27-46ba-4d51-adb5-5365f40da30d","Type":"ContainerDied","Data":"a8a8ccb2c084ae66d812cc1ec4033b06a20329b0cf13dbecaf9788172090a69d"} Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.530251 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-5664bbcc7d-snqwd"] Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.572533 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.576934 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-cbsmf" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.577136 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.577171 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.586605 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5664bbcc7d-snqwd"] Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.589421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-octavia-run\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.589516 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-combined-ca-bundle\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.589662 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.589901 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data-merged\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.590008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-scripts\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.696498 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.702514 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data-merged\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.702890 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-scripts\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.703038 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-octavia-run\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.703126 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-combined-ca-bundle\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.705760 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data-merged\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.705952 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-config-data\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.707593 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-scripts\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.708737 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-octavia-run\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.711846 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c747c0-dca4-4c05-a8ee-9b68b6e2797a-combined-ca-bundle\") pod \"octavia-api-5664bbcc7d-snqwd\" (UID: \"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a\") " pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.832454 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.904737 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916403 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwg44\" (UniqueName: \"kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916457 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916567 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916605 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916668 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run\") pod \"40cf0a27-46ba-4d51-adb5-5365f40da30d\" (UID: \"40cf0a27-46ba-4d51-adb5-5365f40da30d\") " Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916900 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.916956 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.917011 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run" (OuterVolumeSpecName: "var-run") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.917460 4959 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.917478 4959 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.917488 4959 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/40cf0a27-46ba-4d51-adb5-5365f40da30d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.917864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.918145 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts" (OuterVolumeSpecName: "scripts") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:12:19 crc kubenswrapper[4959]: I1003 15:12:19.921599 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44" (OuterVolumeSpecName: "kube-api-access-mwg44") pod "40cf0a27-46ba-4d51-adb5-5365f40da30d" (UID: "40cf0a27-46ba-4d51-adb5-5365f40da30d"). InnerVolumeSpecName "kube-api-access-mwg44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.019723 4959 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.020038 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwg44\" (UniqueName: \"kubernetes.io/projected/40cf0a27-46ba-4d51-adb5-5365f40da30d-kube-api-access-mwg44\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.020051 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40cf0a27-46ba-4d51-adb5-5365f40da30d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.265114 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8nphv-config-hdvr4" event={"ID":"40cf0a27-46ba-4d51-adb5-5365f40da30d","Type":"ContainerDied","Data":"bc445a9fdb53d1cbf8840048c29f348da76fd47a511e4110a072b79fa1d2708d"} Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.265161 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc445a9fdb53d1cbf8840048c29f348da76fd47a511e4110a072b79fa1d2708d" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.265173 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8nphv-config-hdvr4" Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.358757 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-5664bbcc7d-snqwd"] Oct 03 15:12:20 crc kubenswrapper[4959]: W1003 15:12:20.361864 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0c747c0_dca4_4c05_a8ee_9b68b6e2797a.slice/crio-69a13301c6475924ceab09bb42115d7318a366e6b4e6955364d2d9109cbeed22 WatchSource:0}: Error finding container 69a13301c6475924ceab09bb42115d7318a366e6b4e6955364d2d9109cbeed22: Status 404 returned error can't find the container with id 69a13301c6475924ceab09bb42115d7318a366e6b4e6955364d2d9109cbeed22 Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.938045 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8nphv-config-hdvr4"] Oct 03 15:12:20 crc kubenswrapper[4959]: I1003 15:12:20.945147 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8nphv-config-hdvr4"] Oct 03 15:12:21 crc kubenswrapper[4959]: I1003 15:12:21.085669 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8nphv" Oct 03 15:12:21 crc kubenswrapper[4959]: I1003 15:12:21.278845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5664bbcc7d-snqwd" event={"ID":"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a","Type":"ContainerStarted","Data":"69a13301c6475924ceab09bb42115d7318a366e6b4e6955364d2d9109cbeed22"} Oct 03 15:12:21 crc kubenswrapper[4959]: I1003 15:12:21.707967 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40cf0a27-46ba-4d51-adb5-5365f40da30d" path="/var/lib/kubelet/pods/40cf0a27-46ba-4d51-adb5-5365f40da30d/volumes" Oct 03 15:12:29 crc kubenswrapper[4959]: I1003 15:12:29.364989 4959 generic.go:334] "Generic (PLEG): container finished" podID="b0c747c0-dca4-4c05-a8ee-9b68b6e2797a" containerID="96fd6c6ddd9d1c477b5d25dd689dcb05b298d0e97596a3d34fc015e6f670818c" exitCode=0 Oct 03 15:12:29 crc kubenswrapper[4959]: I1003 15:12:29.365229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5664bbcc7d-snqwd" event={"ID":"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a","Type":"ContainerDied","Data":"96fd6c6ddd9d1c477b5d25dd689dcb05b298d0e97596a3d34fc015e6f670818c"} Oct 03 15:12:30 crc kubenswrapper[4959]: I1003 15:12:30.386057 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5664bbcc7d-snqwd" event={"ID":"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a","Type":"ContainerStarted","Data":"28773254897b0dbe2aecc9e8f7835ce4962036966d7feda2ee6394e2bbad2159"} Oct 03 15:12:30 crc kubenswrapper[4959]: I1003 15:12:30.386867 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:30 crc kubenswrapper[4959]: I1003 15:12:30.386887 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-5664bbcc7d-snqwd" event={"ID":"b0c747c0-dca4-4c05-a8ee-9b68b6e2797a","Type":"ContainerStarted","Data":"85d76efa6682e45d10dcb6cfc27eba186f6f931b393026611bbf7bf2cb36d56d"} Oct 03 15:12:30 crc kubenswrapper[4959]: I1003 15:12:30.386925 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:30 crc kubenswrapper[4959]: I1003 15:12:30.412596 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-5664bbcc7d-snqwd" podStartSLOduration=3.579407416 podStartE2EDuration="11.412571751s" podCreationTimestamp="2025-10-03 15:12:19 +0000 UTC" firstStartedPulling="2025-10-03 15:12:20.364102743 +0000 UTC m=+6109.567446160" lastFinishedPulling="2025-10-03 15:12:28.197267078 +0000 UTC m=+6117.400610495" observedRunningTime="2025-10-03 15:12:30.406062981 +0000 UTC m=+6119.609406428" watchObservedRunningTime="2025-10-03 15:12:30.412571751 +0000 UTC m=+6119.615915168" Oct 03 15:12:36 crc kubenswrapper[4959]: I1003 15:12:36.044787 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:12:36 crc kubenswrapper[4959]: I1003 15:12:36.045328 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:12:38 crc kubenswrapper[4959]: I1003 15:12:38.923470 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:38 crc kubenswrapper[4959]: I1003 15:12:38.930701 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-5664bbcc7d-snqwd" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.349692 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-n5vz4"] Oct 03 15:12:39 crc kubenswrapper[4959]: E1003 15:12:39.350288 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40cf0a27-46ba-4d51-adb5-5365f40da30d" containerName="ovn-config" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.350319 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="40cf0a27-46ba-4d51-adb5-5365f40da30d" containerName="ovn-config" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.350633 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="40cf0a27-46ba-4d51-adb5-5365f40da30d" containerName="ovn-config" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.352369 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.354318 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.354847 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.355148 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.369151 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-n5vz4"] Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.428682 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-config-data\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.428782 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/92238071-8adc-41d2-be43-7b92ba7565c7-config-data-merged\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.428863 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-scripts\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.428884 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/92238071-8adc-41d2-be43-7b92ba7565c7-hm-ports\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.530772 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-scripts\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.530816 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/92238071-8adc-41d2-be43-7b92ba7565c7-hm-ports\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.530928 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-config-data\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.530982 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/92238071-8adc-41d2-be43-7b92ba7565c7-config-data-merged\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.531458 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/92238071-8adc-41d2-be43-7b92ba7565c7-config-data-merged\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.532612 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/92238071-8adc-41d2-be43-7b92ba7565c7-hm-ports\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.537158 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-config-data\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.552940 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92238071-8adc-41d2-be43-7b92ba7565c7-scripts\") pod \"octavia-rsyslog-n5vz4\" (UID: \"92238071-8adc-41d2-be43-7b92ba7565c7\") " pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:39 crc kubenswrapper[4959]: I1003 15:12:39.679999 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.257783 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-n5vz4"] Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.332779 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.335025 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.337438 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.344000 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.467959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.468381 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.489738 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-n5vz4" event={"ID":"92238071-8adc-41d2-be43-7b92ba7565c7","Type":"ContainerStarted","Data":"07cfbd9a4c96b684dea40192b5abe78f766bf4b733d7390c0fe8da04798c814b"} Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.570500 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.570591 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.571131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.576707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config\") pod \"octavia-image-upload-59f8cff499-j6ssx\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:40 crc kubenswrapper[4959]: I1003 15:12:40.656475 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:12:41 crc kubenswrapper[4959]: I1003 15:12:41.128153 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:12:41 crc kubenswrapper[4959]: I1003 15:12:41.502166 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerStarted","Data":"94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532"} Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.305260 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-n5bzq"] Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.307108 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.315575 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.316711 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-n5bzq"] Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.435319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.435682 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.435727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.435754 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.521926 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-n5vz4" event={"ID":"92238071-8adc-41d2-be43-7b92ba7565c7","Type":"ContainerStarted","Data":"fcbe40cb12751218b5884c073480889b9e1a695eaf7e2645e787db92271b0e05"} Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.537575 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.537706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.537903 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.537954 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.538717 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.544319 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.546882 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.554533 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts\") pod \"octavia-db-sync-n5bzq\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:42 crc kubenswrapper[4959]: I1003 15:12:42.635978 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:43 crc kubenswrapper[4959]: I1003 15:12:43.576872 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-n5bzq"] Oct 03 15:12:44 crc kubenswrapper[4959]: I1003 15:12:44.557466 4959 generic.go:334] "Generic (PLEG): container finished" podID="f977e88e-bc91-4efc-ac44-686076979802" containerID="a1e57e3b0e723aa9b56db1a4491d4a3cc255c2c37aa2b0192b5d764d3ac1f475" exitCode=0 Oct 03 15:12:44 crc kubenswrapper[4959]: I1003 15:12:44.557621 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-n5bzq" event={"ID":"f977e88e-bc91-4efc-ac44-686076979802","Type":"ContainerDied","Data":"a1e57e3b0e723aa9b56db1a4491d4a3cc255c2c37aa2b0192b5d764d3ac1f475"} Oct 03 15:12:44 crc kubenswrapper[4959]: I1003 15:12:44.558054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-n5bzq" event={"ID":"f977e88e-bc91-4efc-ac44-686076979802","Type":"ContainerStarted","Data":"9948204c6c6a50e4d3f8a550284f95dc99afd0216bec91233392dc99c47b049b"} Oct 03 15:12:44 crc kubenswrapper[4959]: I1003 15:12:44.566770 4959 generic.go:334] "Generic (PLEG): container finished" podID="92238071-8adc-41d2-be43-7b92ba7565c7" containerID="fcbe40cb12751218b5884c073480889b9e1a695eaf7e2645e787db92271b0e05" exitCode=0 Oct 03 15:12:44 crc kubenswrapper[4959]: I1003 15:12:44.566847 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-n5vz4" event={"ID":"92238071-8adc-41d2-be43-7b92ba7565c7","Type":"ContainerDied","Data":"fcbe40cb12751218b5884c073480889b9e1a695eaf7e2645e787db92271b0e05"} Oct 03 15:12:45 crc kubenswrapper[4959]: I1003 15:12:45.577140 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-n5bzq" event={"ID":"f977e88e-bc91-4efc-ac44-686076979802","Type":"ContainerStarted","Data":"f36e0adb5e6ed0a3a1b49a17c906145ac6bfe2a285849bb8a6fea137377e17ef"} Oct 03 15:12:45 crc kubenswrapper[4959]: I1003 15:12:45.600827 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-n5bzq" podStartSLOduration=3.600811037 podStartE2EDuration="3.600811037s" podCreationTimestamp="2025-10-03 15:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:12:45.598752527 +0000 UTC m=+6134.802095934" watchObservedRunningTime="2025-10-03 15:12:45.600811037 +0000 UTC m=+6134.804154454" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.588799 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-n5vz4" event={"ID":"92238071-8adc-41d2-be43-7b92ba7565c7","Type":"ContainerStarted","Data":"ea4d45998fef40f771c759260ba5e6b9ffbf65cac53c86d90b4426bce8d5e152"} Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.591030 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.618909 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-n5vz4" podStartSLOduration=1.705916669 podStartE2EDuration="7.618889432s" podCreationTimestamp="2025-10-03 15:12:39 +0000 UTC" firstStartedPulling="2025-10-03 15:12:40.27034801 +0000 UTC m=+6129.473691437" lastFinishedPulling="2025-10-03 15:12:46.183320783 +0000 UTC m=+6135.386664200" observedRunningTime="2025-10-03 15:12:46.612898094 +0000 UTC m=+6135.816241511" watchObservedRunningTime="2025-10-03 15:12:46.618889432 +0000 UTC m=+6135.822232869" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.708854 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-lqvnf"] Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.710319 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.712273 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.715699 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.718806 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.731335 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-lqvnf"] Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.851418 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88d6d698-a05d-4d67-8844-bef19cb4689e-hm-ports\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.851667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-scripts\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.851946 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-amphora-certs\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.852068 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-combined-ca-bundle\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.852104 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.852130 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data-merged\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954274 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88d6d698-a05d-4d67-8844-bef19cb4689e-hm-ports\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954346 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-scripts\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954390 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-amphora-certs\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954428 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-combined-ca-bundle\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.954466 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data-merged\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.955142 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88d6d698-a05d-4d67-8844-bef19cb4689e-hm-ports\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.955239 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data-merged\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.959395 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-scripts\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.962480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-amphora-certs\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.970467 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-combined-ca-bundle\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:46 crc kubenswrapper[4959]: I1003 15:12:46.972284 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d6d698-a05d-4d67-8844-bef19cb4689e-config-data\") pod \"octavia-healthmanager-lqvnf\" (UID: \"88d6d698-a05d-4d67-8844-bef19cb4689e\") " pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:47 crc kubenswrapper[4959]: I1003 15:12:47.030391 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:47 crc kubenswrapper[4959]: I1003 15:12:47.600676 4959 generic.go:334] "Generic (PLEG): container finished" podID="f977e88e-bc91-4efc-ac44-686076979802" containerID="f36e0adb5e6ed0a3a1b49a17c906145ac6bfe2a285849bb8a6fea137377e17ef" exitCode=0 Oct 03 15:12:47 crc kubenswrapper[4959]: I1003 15:12:47.600714 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-n5bzq" event={"ID":"f977e88e-bc91-4efc-ac44-686076979802","Type":"ContainerDied","Data":"f36e0adb5e6ed0a3a1b49a17c906145ac6bfe2a285849bb8a6fea137377e17ef"} Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.443562 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-zqfhm"] Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.446528 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.449667 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.453467 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.464071 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-zqfhm"] Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.605361 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-hm-ports\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.605720 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.605924 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-scripts\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.606023 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data-merged\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.606156 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-combined-ca-bundle\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.606249 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-amphora-certs\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.710762 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-hm-ports\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.711448 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.711802 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-scripts\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.712024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-hm-ports\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.712626 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data-merged\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.714111 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data-merged\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.714970 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-combined-ca-bundle\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.715799 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-amphora-certs\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.719398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-config-data\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.720576 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-combined-ca-bundle\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.721831 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-scripts\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.727355 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6b0768c0-0cb6-423e-af94-dd2be48aa3a4-amphora-certs\") pod \"octavia-housekeeping-zqfhm\" (UID: \"6b0768c0-0cb6-423e-af94-dd2be48aa3a4\") " pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:49 crc kubenswrapper[4959]: I1003 15:12:49.767667 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.644593 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.649092 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-n5bzq" event={"ID":"f977e88e-bc91-4efc-ac44-686076979802","Type":"ContainerDied","Data":"9948204c6c6a50e4d3f8a550284f95dc99afd0216bec91233392dc99c47b049b"} Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.649125 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9948204c6c6a50e4d3f8a550284f95dc99afd0216bec91233392dc99c47b049b" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.649180 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-n5bzq" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.736307 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data\") pod \"f977e88e-bc91-4efc-ac44-686076979802\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.736423 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged\") pod \"f977e88e-bc91-4efc-ac44-686076979802\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.736641 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts\") pod \"f977e88e-bc91-4efc-ac44-686076979802\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.736731 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") pod \"f977e88e-bc91-4efc-ac44-686076979802\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.740770 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data" (OuterVolumeSpecName: "config-data") pod "f977e88e-bc91-4efc-ac44-686076979802" (UID: "f977e88e-bc91-4efc-ac44-686076979802"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.744643 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts" (OuterVolumeSpecName: "scripts") pod "f977e88e-bc91-4efc-ac44-686076979802" (UID: "f977e88e-bc91-4efc-ac44-686076979802"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:12:50 crc kubenswrapper[4959]: E1003 15:12:50.767452 4959 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle podName:f977e88e-bc91-4efc-ac44-686076979802 nodeName:}" failed. No retries permitted until 2025-10-03 15:12:51.267425875 +0000 UTC m=+6140.470769302 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle") pod "f977e88e-bc91-4efc-ac44-686076979802" (UID: "f977e88e-bc91-4efc-ac44-686076979802") : error deleting /var/lib/kubelet/pods/f977e88e-bc91-4efc-ac44-686076979802/volume-subpaths: remove /var/lib/kubelet/pods/f977e88e-bc91-4efc-ac44-686076979802/volume-subpaths: no such file or directory Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.767724 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "f977e88e-bc91-4efc-ac44-686076979802" (UID: "f977e88e-bc91-4efc-ac44-686076979802"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.839835 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.839872 4959 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f977e88e-bc91-4efc-ac44-686076979802-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:50 crc kubenswrapper[4959]: I1003 15:12:50.839884 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.349461 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") pod \"f977e88e-bc91-4efc-ac44-686076979802\" (UID: \"f977e88e-bc91-4efc-ac44-686076979802\") " Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.364795 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f977e88e-bc91-4efc-ac44-686076979802" (UID: "f977e88e-bc91-4efc-ac44-686076979802"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.451729 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f977e88e-bc91-4efc-ac44-686076979802-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.628747 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-zqfhm"] Oct 03 15:12:51 crc kubenswrapper[4959]: W1003 15:12:51.634153 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b0768c0_0cb6_423e_af94_dd2be48aa3a4.slice/crio-55ddb7c47e55a9fdbdf9d11c50c363bc2030d239643dfae3336476399fe87405 WatchSource:0}: Error finding container 55ddb7c47e55a9fdbdf9d11c50c363bc2030d239643dfae3336476399fe87405: Status 404 returned error can't find the container with id 55ddb7c47e55a9fdbdf9d11c50c363bc2030d239643dfae3336476399fe87405 Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.662670 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerStarted","Data":"1f8822c433f955a6b5e556935b81b14210d9a66e3946f4b97362a752a07c9720"} Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.664677 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zqfhm" event={"ID":"6b0768c0-0cb6-423e-af94-dd2be48aa3a4","Type":"ContainerStarted","Data":"55ddb7c47e55a9fdbdf9d11c50c363bc2030d239643dfae3336476399fe87405"} Oct 03 15:12:51 crc kubenswrapper[4959]: I1003 15:12:51.765704 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-lqvnf"] Oct 03 15:12:52 crc kubenswrapper[4959]: I1003 15:12:52.674856 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-lqvnf" event={"ID":"88d6d698-a05d-4d67-8844-bef19cb4689e","Type":"ContainerStarted","Data":"13992c8fd7a82df250122b45e01960cbb4043c5d256e264f1b9a30f4af82746b"} Oct 03 15:12:52 crc kubenswrapper[4959]: I1003 15:12:52.675164 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-lqvnf" event={"ID":"88d6d698-a05d-4d67-8844-bef19cb4689e","Type":"ContainerStarted","Data":"50f927a36ced7254b3f3cfdae11f576d519ef648c332b7f495c993bac81a712d"} Oct 03 15:12:52 crc kubenswrapper[4959]: I1003 15:12:52.677665 4959 generic.go:334] "Generic (PLEG): container finished" podID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerID="1f8822c433f955a6b5e556935b81b14210d9a66e3946f4b97362a752a07c9720" exitCode=0 Oct 03 15:12:52 crc kubenswrapper[4959]: I1003 15:12:52.677717 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerDied","Data":"1f8822c433f955a6b5e556935b81b14210d9a66e3946f4b97362a752a07c9720"} Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.717790 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-n5vz4" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.735154 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-nc45h"] Oct 03 15:12:54 crc kubenswrapper[4959]: E1003 15:12:54.736285 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f977e88e-bc91-4efc-ac44-686076979802" containerName="init" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.736312 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f977e88e-bc91-4efc-ac44-686076979802" containerName="init" Oct 03 15:12:54 crc kubenswrapper[4959]: E1003 15:12:54.736340 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f977e88e-bc91-4efc-ac44-686076979802" containerName="octavia-db-sync" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.736351 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f977e88e-bc91-4efc-ac44-686076979802" containerName="octavia-db-sync" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.736936 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f977e88e-bc91-4efc-ac44-686076979802" containerName="octavia-db-sync" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.739571 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.755004 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.755147 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.761668 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerStarted","Data":"c699d13d4e4a46ff952c4d3ef26d26b4c380ef4cdc4a7dbac54932a1716aec09"} Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.763228 4959 generic.go:334] "Generic (PLEG): container finished" podID="6b0768c0-0cb6-423e-af94-dd2be48aa3a4" containerID="8481fa728ea80075f6ee5a6c5c099cf065236bd4d9d3d5bcb257550ab8cee887" exitCode=0 Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.764590 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zqfhm" event={"ID":"6b0768c0-0cb6-423e-af94-dd2be48aa3a4","Type":"ContainerDied","Data":"8481fa728ea80075f6ee5a6c5c099cf065236bd4d9d3d5bcb257550ab8cee887"} Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.793971 4959 generic.go:334] "Generic (PLEG): container finished" podID="88d6d698-a05d-4d67-8844-bef19cb4689e" containerID="13992c8fd7a82df250122b45e01960cbb4043c5d256e264f1b9a30f4af82746b" exitCode=0 Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.794015 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-lqvnf" event={"ID":"88d6d698-a05d-4d67-8844-bef19cb4689e","Type":"ContainerDied","Data":"13992c8fd7a82df250122b45e01960cbb4043c5d256e264f1b9a30f4af82746b"} Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.801558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-nc45h"] Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.876750 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" podStartSLOduration=1.949523984 podStartE2EDuration="14.876733738s" podCreationTimestamp="2025-10-03 15:12:40 +0000 UTC" firstStartedPulling="2025-10-03 15:12:41.134821976 +0000 UTC m=+6130.338165393" lastFinishedPulling="2025-10-03 15:12:54.06203173 +0000 UTC m=+6143.265375147" observedRunningTime="2025-10-03 15:12:54.866060976 +0000 UTC m=+6144.069404393" watchObservedRunningTime="2025-10-03 15:12:54.876733738 +0000 UTC m=+6144.080077155" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.920798 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-config-data\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.920980 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26e1c5e1-5763-4b18-8410-19f956516874-config-data-merged\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.921061 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-amphora-certs\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.921109 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-scripts\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.921130 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26e1c5e1-5763-4b18-8410-19f956516874-hm-ports\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:54 crc kubenswrapper[4959]: I1003 15:12:54.921158 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-combined-ca-bundle\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.022794 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26e1c5e1-5763-4b18-8410-19f956516874-config-data-merged\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023096 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-amphora-certs\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023127 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-scripts\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023146 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26e1c5e1-5763-4b18-8410-19f956516874-hm-ports\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023164 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-combined-ca-bundle\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023215 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-config-data\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.023490 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26e1c5e1-5763-4b18-8410-19f956516874-config-data-merged\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.027274 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26e1c5e1-5763-4b18-8410-19f956516874-hm-ports\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.029477 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-combined-ca-bundle\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.031675 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-amphora-certs\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.043454 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-config-data\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.043558 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e1c5e1-5763-4b18-8410-19f956516874-scripts\") pod \"octavia-worker-nc45h\" (UID: \"26e1c5e1-5763-4b18-8410-19f956516874\") " pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.252099 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-nc45h" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.824221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-zqfhm" event={"ID":"6b0768c0-0cb6-423e-af94-dd2be48aa3a4","Type":"ContainerStarted","Data":"2a6094849a9de347bdf491a6017878c6cffce60fc6857839858aa9b9b8157ea2"} Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.824762 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.826991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-lqvnf" event={"ID":"88d6d698-a05d-4d67-8844-bef19cb4689e","Type":"ContainerStarted","Data":"2b0da3f162b4df4037c95e24ab516567b7a3609aed4d48651bf3060eee1be5e8"} Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.827921 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.860826 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-zqfhm" podStartSLOduration=5.191383665 podStartE2EDuration="6.860796659s" podCreationTimestamp="2025-10-03 15:12:49 +0000 UTC" firstStartedPulling="2025-10-03 15:12:51.639795473 +0000 UTC m=+6140.843138890" lastFinishedPulling="2025-10-03 15:12:53.309208457 +0000 UTC m=+6142.512551884" observedRunningTime="2025-10-03 15:12:55.850138228 +0000 UTC m=+6145.053481655" watchObservedRunningTime="2025-10-03 15:12:55.860796659 +0000 UTC m=+6145.064140066" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.906219 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-lqvnf" podStartSLOduration=9.90616605 podStartE2EDuration="9.90616605s" podCreationTimestamp="2025-10-03 15:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:12:55.886142469 +0000 UTC m=+6145.089485896" watchObservedRunningTime="2025-10-03 15:12:55.90616605 +0000 UTC m=+6145.109509467" Oct 03 15:12:55 crc kubenswrapper[4959]: I1003 15:12:55.920469 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-nc45h"] Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.125557 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.127500 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.143866 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.254428 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.254528 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.254648 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zwhw\" (UniqueName: \"kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.356117 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zwhw\" (UniqueName: \"kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.356229 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.356330 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.356912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.357007 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.379528 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zwhw\" (UniqueName: \"kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw\") pod \"redhat-operators-bqmj5\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.448859 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.721590 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.723854 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.739978 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.843091 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-nc45h" event={"ID":"26e1c5e1-5763-4b18-8410-19f956516874","Type":"ContainerStarted","Data":"94296f8b5c7bcad921703d98edf4a9f167aea9388088735de35a25c365f599b8"} Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.869729 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58k4f\" (UniqueName: \"kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.869904 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.869988 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.972123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.972238 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.972357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58k4f\" (UniqueName: \"kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.972666 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.973162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:56 crc kubenswrapper[4959]: I1003 15:12:56.992773 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58k4f\" (UniqueName: \"kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f\") pod \"community-operators-n7wdn\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:57 crc kubenswrapper[4959]: I1003 15:12:57.060908 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:12:57 crc kubenswrapper[4959]: I1003 15:12:57.210263 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:12:57 crc kubenswrapper[4959]: I1003 15:12:57.856390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerStarted","Data":"521d520901deb230618fc7041d90e48653469228682628a59832c4ffc8c5435f"} Oct 03 15:12:57 crc kubenswrapper[4959]: I1003 15:12:57.884859 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:12:58 crc kubenswrapper[4959]: I1003 15:12:58.879392 4959 generic.go:334] "Generic (PLEG): container finished" podID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerID="725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b" exitCode=0 Oct 03 15:12:58 crc kubenswrapper[4959]: I1003 15:12:58.879489 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerDied","Data":"725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b"} Oct 03 15:12:58 crc kubenswrapper[4959]: I1003 15:12:58.885613 4959 generic.go:334] "Generic (PLEG): container finished" podID="a6a61e2c-89f9-492a-895f-810c1a107551" containerID="803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0" exitCode=0 Oct 03 15:12:58 crc kubenswrapper[4959]: I1003 15:12:58.885678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerDied","Data":"803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0"} Oct 03 15:12:58 crc kubenswrapper[4959]: I1003 15:12:58.885750 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerStarted","Data":"f9321d5c113ec827d124fd4ee7a9857a8e732c609b54b989b45226f7146b875e"} Oct 03 15:13:00 crc kubenswrapper[4959]: I1003 15:13:00.902399 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerStarted","Data":"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37"} Oct 03 15:13:00 crc kubenswrapper[4959]: I1003 15:13:00.904516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerStarted","Data":"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f"} Oct 03 15:13:01 crc kubenswrapper[4959]: I1003 15:13:01.914442 4959 generic.go:334] "Generic (PLEG): container finished" podID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerID="19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37" exitCode=0 Oct 03 15:13:01 crc kubenswrapper[4959]: I1003 15:13:01.914508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerDied","Data":"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37"} Oct 03 15:13:01 crc kubenswrapper[4959]: I1003 15:13:01.917498 4959 generic.go:334] "Generic (PLEG): container finished" podID="a6a61e2c-89f9-492a-895f-810c1a107551" containerID="5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f" exitCode=0 Oct 03 15:13:01 crc kubenswrapper[4959]: I1003 15:13:01.917572 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerDied","Data":"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f"} Oct 03 15:13:01 crc kubenswrapper[4959]: I1003 15:13:01.920309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-nc45h" event={"ID":"26e1c5e1-5763-4b18-8410-19f956516874","Type":"ContainerStarted","Data":"ef3940583ac64f4df5e43a7985533716460040ac853b342d6c8fb7034660f4c3"} Oct 03 15:13:02 crc kubenswrapper[4959]: I1003 15:13:02.065891 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-lqvnf" Oct 03 15:13:03 crc kubenswrapper[4959]: I1003 15:13:03.939875 4959 generic.go:334] "Generic (PLEG): container finished" podID="26e1c5e1-5763-4b18-8410-19f956516874" containerID="ef3940583ac64f4df5e43a7985533716460040ac853b342d6c8fb7034660f4c3" exitCode=0 Oct 03 15:13:03 crc kubenswrapper[4959]: I1003 15:13:03.940118 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-nc45h" event={"ID":"26e1c5e1-5763-4b18-8410-19f956516874","Type":"ContainerDied","Data":"ef3940583ac64f4df5e43a7985533716460040ac853b342d6c8fb7034660f4c3"} Oct 03 15:13:03 crc kubenswrapper[4959]: I1003 15:13:03.949385 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerStarted","Data":"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540"} Oct 03 15:13:03 crc kubenswrapper[4959]: I1003 15:13:03.958653 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerStarted","Data":"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2"} Oct 03 15:13:04 crc kubenswrapper[4959]: I1003 15:13:04.004699 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bqmj5" podStartSLOduration=3.806323563 podStartE2EDuration="8.004681556s" podCreationTimestamp="2025-10-03 15:12:56 +0000 UTC" firstStartedPulling="2025-10-03 15:12:58.882485125 +0000 UTC m=+6148.085828542" lastFinishedPulling="2025-10-03 15:13:03.080843118 +0000 UTC m=+6152.284186535" observedRunningTime="2025-10-03 15:13:03.994540998 +0000 UTC m=+6153.197884455" watchObservedRunningTime="2025-10-03 15:13:04.004681556 +0000 UTC m=+6153.208024973" Oct 03 15:13:04 crc kubenswrapper[4959]: I1003 15:13:04.798008 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-zqfhm" Oct 03 15:13:04 crc kubenswrapper[4959]: I1003 15:13:04.972141 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-nc45h" event={"ID":"26e1c5e1-5763-4b18-8410-19f956516874","Type":"ContainerStarted","Data":"08c85063afd78d11c8a6a98e9b1f4d2e7ef4f9141cc298f92299e9fe47bce4f2"} Oct 03 15:13:04 crc kubenswrapper[4959]: I1003 15:13:04.993948 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n7wdn" podStartSLOduration=4.911766326 podStartE2EDuration="8.993928765s" podCreationTimestamp="2025-10-03 15:12:56 +0000 UTC" firstStartedPulling="2025-10-03 15:12:58.887416666 +0000 UTC m=+6148.090760083" lastFinishedPulling="2025-10-03 15:13:02.969579105 +0000 UTC m=+6152.172922522" observedRunningTime="2025-10-03 15:13:04.990819189 +0000 UTC m=+6154.194162616" watchObservedRunningTime="2025-10-03 15:13:04.993928765 +0000 UTC m=+6154.197272182" Oct 03 15:13:05 crc kubenswrapper[4959]: I1003 15:13:05.015681 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-nc45h" podStartSLOduration=6.668907474 podStartE2EDuration="11.015661187s" podCreationTimestamp="2025-10-03 15:12:54 +0000 UTC" firstStartedPulling="2025-10-03 15:12:55.924658662 +0000 UTC m=+6145.128002079" lastFinishedPulling="2025-10-03 15:13:00.271412375 +0000 UTC m=+6149.474755792" observedRunningTime="2025-10-03 15:13:05.01249688 +0000 UTC m=+6154.215840297" watchObservedRunningTime="2025-10-03 15:13:05.015661187 +0000 UTC m=+6154.219004604" Oct 03 15:13:06 crc kubenswrapper[4959]: I1003 15:13:06.044236 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:13:06 crc kubenswrapper[4959]: I1003 15:13:06.044561 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:13:06 crc kubenswrapper[4959]: I1003 15:13:06.449819 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:06 crc kubenswrapper[4959]: I1003 15:13:06.449880 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:07 crc kubenswrapper[4959]: I1003 15:13:07.062146 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:07 crc kubenswrapper[4959]: I1003 15:13:07.062541 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:07 crc kubenswrapper[4959]: I1003 15:13:07.114142 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:07 crc kubenswrapper[4959]: I1003 15:13:07.525330 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bqmj5" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="registry-server" probeResult="failure" output=< Oct 03 15:13:07 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:13:07 crc kubenswrapper[4959]: > Oct 03 15:13:08 crc kubenswrapper[4959]: I1003 15:13:08.058583 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:08 crc kubenswrapper[4959]: I1003 15:13:08.714250 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.023237 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n7wdn" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="registry-server" containerID="cri-o://6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2" gracePeriod=2 Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.252854 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-nc45h" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.294554 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-nc45h" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.553820 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.676563 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58k4f\" (UniqueName: \"kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f\") pod \"a6a61e2c-89f9-492a-895f-810c1a107551\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.676728 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities\") pod \"a6a61e2c-89f9-492a-895f-810c1a107551\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.676759 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content\") pod \"a6a61e2c-89f9-492a-895f-810c1a107551\" (UID: \"a6a61e2c-89f9-492a-895f-810c1a107551\") " Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.677728 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities" (OuterVolumeSpecName: "utilities") pod "a6a61e2c-89f9-492a-895f-810c1a107551" (UID: "a6a61e2c-89f9-492a-895f-810c1a107551"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.678266 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.686863 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f" (OuterVolumeSpecName: "kube-api-access-58k4f") pod "a6a61e2c-89f9-492a-895f-810c1a107551" (UID: "a6a61e2c-89f9-492a-895f-810c1a107551"). InnerVolumeSpecName "kube-api-access-58k4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.735906 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6a61e2c-89f9-492a-895f-810c1a107551" (UID: "a6a61e2c-89f9-492a-895f-810c1a107551"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.780599 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58k4f\" (UniqueName: \"kubernetes.io/projected/a6a61e2c-89f9-492a-895f-810c1a107551-kube-api-access-58k4f\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:10 crc kubenswrapper[4959]: I1003 15:13:10.780626 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6a61e2c-89f9-492a-895f-810c1a107551-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.037349 4959 generic.go:334] "Generic (PLEG): container finished" podID="a6a61e2c-89f9-492a-895f-810c1a107551" containerID="6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2" exitCode=0 Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.038464 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerDied","Data":"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2"} Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.038524 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7wdn" event={"ID":"a6a61e2c-89f9-492a-895f-810c1a107551","Type":"ContainerDied","Data":"f9321d5c113ec827d124fd4ee7a9857a8e732c609b54b989b45226f7146b875e"} Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.038550 4959 scope.go:117] "RemoveContainer" containerID="6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.041420 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7wdn" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.073660 4959 scope.go:117] "RemoveContainer" containerID="5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.112526 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.122724 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n7wdn"] Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.131859 4959 scope.go:117] "RemoveContainer" containerID="803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.162020 4959 scope.go:117] "RemoveContainer" containerID="6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2" Oct 03 15:13:11 crc kubenswrapper[4959]: E1003 15:13:11.162542 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2\": container with ID starting with 6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2 not found: ID does not exist" containerID="6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.162651 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2"} err="failed to get container status \"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2\": rpc error: code = NotFound desc = could not find container \"6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2\": container with ID starting with 6f9099f1cd1a7e8d85b11bf82e427674667afa8781d2a25c93e45063585c9ae2 not found: ID does not exist" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.162740 4959 scope.go:117] "RemoveContainer" containerID="5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f" Oct 03 15:13:11 crc kubenswrapper[4959]: E1003 15:13:11.163376 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f\": container with ID starting with 5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f not found: ID does not exist" containerID="5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.163439 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f"} err="failed to get container status \"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f\": rpc error: code = NotFound desc = could not find container \"5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f\": container with ID starting with 5836520c46d312a3f379b16e5cad85dac1e288c8c923e53a8747604caf1bd43f not found: ID does not exist" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.163478 4959 scope.go:117] "RemoveContainer" containerID="803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0" Oct 03 15:13:11 crc kubenswrapper[4959]: E1003 15:13:11.163880 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0\": container with ID starting with 803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0 not found: ID does not exist" containerID="803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.163948 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0"} err="failed to get container status \"803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0\": rpc error: code = NotFound desc = could not find container \"803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0\": container with ID starting with 803ad16629e3cf069a83f61122cbc383d2348836e1c961feea8cd56d09f720d0 not found: ID does not exist" Oct 03 15:13:11 crc kubenswrapper[4959]: E1003 15:13:11.260265 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6a61e2c_89f9_492a_895f_810c1a107551.slice/crio-f9321d5c113ec827d124fd4ee7a9857a8e732c609b54b989b45226f7146b875e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6a61e2c_89f9_492a_895f_810c1a107551.slice\": RecentStats: unable to find data in memory cache]" Oct 03 15:13:11 crc kubenswrapper[4959]: I1003 15:13:11.699736 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" path="/var/lib/kubelet/pods/a6a61e2c-89f9-492a-895f-810c1a107551/volumes" Oct 03 15:13:16 crc kubenswrapper[4959]: I1003 15:13:16.544771 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:16 crc kubenswrapper[4959]: I1003 15:13:16.621381 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:16 crc kubenswrapper[4959]: I1003 15:13:16.788544 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.113322 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bqmj5" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="registry-server" containerID="cri-o://e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540" gracePeriod=2 Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.588647 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.653633 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content\") pod \"8f8f8128-73ec-4fe0-b83d-911417330f0b\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.653778 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zwhw\" (UniqueName: \"kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw\") pod \"8f8f8128-73ec-4fe0-b83d-911417330f0b\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.653868 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities\") pod \"8f8f8128-73ec-4fe0-b83d-911417330f0b\" (UID: \"8f8f8128-73ec-4fe0-b83d-911417330f0b\") " Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.659437 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities" (OuterVolumeSpecName: "utilities") pod "8f8f8128-73ec-4fe0-b83d-911417330f0b" (UID: "8f8f8128-73ec-4fe0-b83d-911417330f0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.668404 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw" (OuterVolumeSpecName: "kube-api-access-5zwhw") pod "8f8f8128-73ec-4fe0-b83d-911417330f0b" (UID: "8f8f8128-73ec-4fe0-b83d-911417330f0b"). InnerVolumeSpecName "kube-api-access-5zwhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.731460 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f8f8128-73ec-4fe0-b83d-911417330f0b" (UID: "8f8f8128-73ec-4fe0-b83d-911417330f0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.760771 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.761001 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zwhw\" (UniqueName: \"kubernetes.io/projected/8f8f8128-73ec-4fe0-b83d-911417330f0b-kube-api-access-5zwhw\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:18 crc kubenswrapper[4959]: I1003 15:13:18.761080 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8f8128-73ec-4fe0-b83d-911417330f0b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.039341 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8l4kz"] Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.048084 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8l4kz"] Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.128805 4959 generic.go:334] "Generic (PLEG): container finished" podID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerID="e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540" exitCode=0 Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.128857 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerDied","Data":"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540"} Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.128888 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bqmj5" event={"ID":"8f8f8128-73ec-4fe0-b83d-911417330f0b","Type":"ContainerDied","Data":"521d520901deb230618fc7041d90e48653469228682628a59832c4ffc8c5435f"} Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.128907 4959 scope.go:117] "RemoveContainer" containerID="e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.129937 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bqmj5" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.156063 4959 scope.go:117] "RemoveContainer" containerID="19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.166934 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.178124 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bqmj5"] Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.202458 4959 scope.go:117] "RemoveContainer" containerID="725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.226006 4959 scope.go:117] "RemoveContainer" containerID="e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540" Oct 03 15:13:19 crc kubenswrapper[4959]: E1003 15:13:19.226563 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540\": container with ID starting with e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540 not found: ID does not exist" containerID="e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.226701 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540"} err="failed to get container status \"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540\": rpc error: code = NotFound desc = could not find container \"e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540\": container with ID starting with e33c67427aac4ee1c673b6cfd8147f6d85ec541913dc78cda052e76af8a4e540 not found: ID does not exist" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.226790 4959 scope.go:117] "RemoveContainer" containerID="19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37" Oct 03 15:13:19 crc kubenswrapper[4959]: E1003 15:13:19.227259 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37\": container with ID starting with 19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37 not found: ID does not exist" containerID="19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.227339 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37"} err="failed to get container status \"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37\": rpc error: code = NotFound desc = could not find container \"19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37\": container with ID starting with 19e445343f15dfee5eb29b249fb2fa5eb44f2987b6ecbbcd8817acc2124a0d37 not found: ID does not exist" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.227401 4959 scope.go:117] "RemoveContainer" containerID="725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b" Oct 03 15:13:19 crc kubenswrapper[4959]: E1003 15:13:19.227729 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b\": container with ID starting with 725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b not found: ID does not exist" containerID="725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.227810 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b"} err="failed to get container status \"725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b\": rpc error: code = NotFound desc = could not find container \"725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b\": container with ID starting with 725f8a66d56bd5f935cdd8a15f6c335c042c3d6ae4e4db488837df607860d42b not found: ID does not exist" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.705628 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d83140-12ff-4a67-83df-5d645515c275" path="/var/lib/kubelet/pods/89d83140-12ff-4a67-83df-5d645515c275/volumes" Oct 03 15:13:19 crc kubenswrapper[4959]: I1003 15:13:19.706825 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" path="/var/lib/kubelet/pods/8f8f8128-73ec-4fe0-b83d-911417330f0b/volumes" Oct 03 15:13:29 crc kubenswrapper[4959]: I1003 15:13:29.026597 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a79-account-create-pqb8q"] Oct 03 15:13:29 crc kubenswrapper[4959]: I1003 15:13:29.035363 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8a79-account-create-pqb8q"] Oct 03 15:13:29 crc kubenswrapper[4959]: I1003 15:13:29.696731 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76be3406-5b1b-47a6-990e-1c28f52e9ada" path="/var/lib/kubelet/pods/76be3406-5b1b-47a6-990e-1c28f52e9ada/volumes" Oct 03 15:13:30 crc kubenswrapper[4959]: I1003 15:13:30.826758 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:13:30 crc kubenswrapper[4959]: I1003 15:13:30.827427 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="octavia-amphora-httpd" containerID="cri-o://c699d13d4e4a46ff952c4d3ef26d26b4c380ef4cdc4a7dbac54932a1716aec09" gracePeriod=30 Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.249962 4959 generic.go:334] "Generic (PLEG): container finished" podID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerID="c699d13d4e4a46ff952c4d3ef26d26b4c380ef4cdc4a7dbac54932a1716aec09" exitCode=0 Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.250065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerDied","Data":"c699d13d4e4a46ff952c4d3ef26d26b4c380ef4cdc4a7dbac54932a1716aec09"} Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.513765 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.622852 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config\") pod \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.623087 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image\") pod \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\" (UID: \"99a72d18-2ef4-45d8-a807-eb7b3162bebf\") " Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.673358 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "99a72d18-2ef4-45d8-a807-eb7b3162bebf" (UID: "99a72d18-2ef4-45d8-a807-eb7b3162bebf"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.676241 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "99a72d18-2ef4-45d8-a807-eb7b3162bebf" (UID: "99a72d18-2ef4-45d8-a807-eb7b3162bebf"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.729714 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/99a72d18-2ef4-45d8-a807-eb7b3162bebf-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:31 crc kubenswrapper[4959]: I1003 15:13:31.729781 4959 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/99a72d18-2ef4-45d8-a807-eb7b3162bebf-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.261129 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" event={"ID":"99a72d18-2ef4-45d8-a807-eb7b3162bebf","Type":"ContainerDied","Data":"94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532"} Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.261272 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-j6ssx" Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.261431 4959 scope.go:117] "RemoveContainer" containerID="c699d13d4e4a46ff952c4d3ef26d26b4c380ef4cdc4a7dbac54932a1716aec09" Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.291742 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.296241 4959 scope.go:117] "RemoveContainer" containerID="1f8822c433f955a6b5e556935b81b14210d9a66e3946f4b97362a752a07c9720" Oct 03 15:13:32 crc kubenswrapper[4959]: I1003 15:13:32.300242 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-j6ssx"] Oct 03 15:13:33 crc kubenswrapper[4959]: I1003 15:13:33.703127 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" path="/var/lib/kubelet/pods/99a72d18-2ef4-45d8-a807-eb7b3162bebf/volumes" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.037778 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-prmj8"] Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.044895 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.044999 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.045129 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.046621 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.046763 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a" gracePeriod=600 Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.054357 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-prmj8"] Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.103423 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-zt5s2"] Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104038 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="extract-content" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104179 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="extract-content" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104261 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="extract-utilities" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104308 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="extract-utilities" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104364 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="extract-utilities" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104408 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="extract-utilities" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104471 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="init" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104521 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="init" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104572 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104616 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104680 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="octavia-amphora-httpd" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104778 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="octavia-amphora-httpd" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104846 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="extract-content" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104895 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="extract-content" Oct 03 15:13:36 crc kubenswrapper[4959]: E1003 15:13:36.104947 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.104994 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.105253 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a61e2c-89f9-492a-895f-810c1a107551" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.105365 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8f8128-73ec-4fe0-b83d-911417330f0b" containerName="registry-server" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.105499 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a72d18-2ef4-45d8-a807-eb7b3162bebf" containerName="octavia-amphora-httpd" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.106668 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.109887 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.141820 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-zt5s2"] Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.222008 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b24d93bb-3ffd-43c6-9c19-377d84d71cde-httpd-config\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.222406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b24d93bb-3ffd-43c6-9c19-377d84d71cde-amphora-image\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.319643 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a" exitCode=0 Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.319701 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a"} Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.319748 4959 scope.go:117] "RemoveContainer" containerID="0b4f400b8b9afe786993a71bfdee1e153ddf2388d0cc2b7746a7854c848c4505" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.324387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b24d93bb-3ffd-43c6-9c19-377d84d71cde-httpd-config\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.324549 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b24d93bb-3ffd-43c6-9c19-377d84d71cde-amphora-image\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.324938 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b24d93bb-3ffd-43c6-9c19-377d84d71cde-amphora-image\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.334902 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b24d93bb-3ffd-43c6-9c19-377d84d71cde-httpd-config\") pod \"octavia-image-upload-59f8cff499-zt5s2\" (UID: \"b24d93bb-3ffd-43c6-9c19-377d84d71cde\") " pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.441525 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" Oct 03 15:13:36 crc kubenswrapper[4959]: I1003 15:13:36.874401 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-zt5s2"] Oct 03 15:13:36 crc kubenswrapper[4959]: W1003 15:13:36.875436 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24d93bb_3ffd_43c6_9c19_377d84d71cde.slice/crio-5365432b5566a866693f05e1149f086f94011b6cd80286e7802aeec74890e3d7 WatchSource:0}: Error finding container 5365432b5566a866693f05e1149f086f94011b6cd80286e7802aeec74890e3d7: Status 404 returned error can't find the container with id 5365432b5566a866693f05e1149f086f94011b6cd80286e7802aeec74890e3d7 Oct 03 15:13:37 crc kubenswrapper[4959]: I1003 15:13:37.330719 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" event={"ID":"b24d93bb-3ffd-43c6-9c19-377d84d71cde","Type":"ContainerStarted","Data":"5365432b5566a866693f05e1149f086f94011b6cd80286e7802aeec74890e3d7"} Oct 03 15:13:37 crc kubenswrapper[4959]: I1003 15:13:37.333035 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176"} Oct 03 15:13:37 crc kubenswrapper[4959]: I1003 15:13:37.699572 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7acf8ef-a0b7-4f3a-a180-0c3f56580328" path="/var/lib/kubelet/pods/a7acf8ef-a0b7-4f3a-a180-0c3f56580328/volumes" Oct 03 15:13:38 crc kubenswrapper[4959]: I1003 15:13:38.344121 4959 generic.go:334] "Generic (PLEG): container finished" podID="b24d93bb-3ffd-43c6-9c19-377d84d71cde" containerID="8f301758364f85191f7b746f6afa09501d564e121b04a36db9724e78a6567dac" exitCode=0 Oct 03 15:13:38 crc kubenswrapper[4959]: I1003 15:13:38.344163 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" event={"ID":"b24d93bb-3ffd-43c6-9c19-377d84d71cde","Type":"ContainerDied","Data":"8f301758364f85191f7b746f6afa09501d564e121b04a36db9724e78a6567dac"} Oct 03 15:13:40 crc kubenswrapper[4959]: I1003 15:13:40.364065 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" event={"ID":"b24d93bb-3ffd-43c6-9c19-377d84d71cde","Type":"ContainerStarted","Data":"a8a2b76c75ac647bacf900a5b9b465ec960f71a00559d93dc6400e62b0e464c8"} Oct 03 15:13:40 crc kubenswrapper[4959]: I1003 15:13:40.384347 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-zt5s2" podStartSLOduration=1.4359229359999999 podStartE2EDuration="4.384320019s" podCreationTimestamp="2025-10-03 15:13:36 +0000 UTC" firstStartedPulling="2025-10-03 15:13:36.87873072 +0000 UTC m=+6186.082074137" lastFinishedPulling="2025-10-03 15:13:39.827127803 +0000 UTC m=+6189.030471220" observedRunningTime="2025-10-03 15:13:40.379384127 +0000 UTC m=+6189.582727554" watchObservedRunningTime="2025-10-03 15:13:40.384320019 +0000 UTC m=+6189.587663436" Oct 03 15:13:42 crc kubenswrapper[4959]: E1003 15:13:42.070226 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice/crio-94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532\": RecentStats: unable to find data in memory cache]" Oct 03 15:13:52 crc kubenswrapper[4959]: E1003 15:13:52.321499 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice/crio-94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532\": RecentStats: unable to find data in memory cache]" Oct 03 15:14:02 crc kubenswrapper[4959]: E1003 15:14:02.581136 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice/crio-94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532\": RecentStats: unable to find data in memory cache]" Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.103001 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-npfx5"] Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.113330 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-npfx5"] Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.783269 4959 scope.go:117] "RemoveContainer" containerID="eed0267e026a834788e77b5abf7f511f02ca3f55bfa538c2ab88020c216a8a22" Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.815837 4959 scope.go:117] "RemoveContainer" containerID="04f3cb5cf8e17e03130963fa7fb750bee7ad94650c21e349f56637d1c63e0fd7" Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.862501 4959 scope.go:117] "RemoveContainer" containerID="ee9ddde554b4e72117583d47ba1873bcfe4b4eaedebf8cd77a1c50d7d26c1e4a" Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.897295 4959 scope.go:117] "RemoveContainer" containerID="367703c6fd0a2b36361ffd3982b08910634276352fc8f1f7b635b00e716286ad" Oct 03 15:14:06 crc kubenswrapper[4959]: I1003 15:14:06.944316 4959 scope.go:117] "RemoveContainer" containerID="5876ccf01e94c8c4a7f84a47ca061eadd20ab74383c3202dd5510873188c2ea6" Oct 03 15:14:07 crc kubenswrapper[4959]: I1003 15:14:07.698146 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e14283b9-fe65-49a2-ac39-43429bb24a4e" path="/var/lib/kubelet/pods/e14283b9-fe65-49a2-ac39-43429bb24a4e/volumes" Oct 03 15:14:12 crc kubenswrapper[4959]: E1003 15:14:12.847285 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice/crio-94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532\": RecentStats: unable to find data in memory cache]" Oct 03 15:14:16 crc kubenswrapper[4959]: I1003 15:14:16.026328 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6131-account-create-gmvbg"] Oct 03 15:14:16 crc kubenswrapper[4959]: I1003 15:14:16.033463 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6131-account-create-gmvbg"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.231463 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.234215 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.237087 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.237694 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.237961 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-5l8jr" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.239329 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.264279 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.300775 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.311404 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-log" containerID="cri-o://b803fd51295e215eae9411738e10fe8923b5148585e6b75c37e71b66d6a252d4" gracePeriod=30 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.311897 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-httpd" containerID="cri-o://43f8f5ac309a5bb879d4e686502c58a47a358376d14d8d63685b476f5fa43540" gracePeriod=30 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.337101 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.337162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfwf\" (UniqueName: \"kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.337319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.337458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.338515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.359356 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.367432 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.409177 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.421296 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.421657 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-log" containerID="cri-o://b50cf27ea1530345a2dfa9e9f67ae26f85c66ad093aebeefc2e5dc15adb45c0c" gracePeriod=30 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.422254 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-httpd" containerID="cri-o://2bbff60e17c67bb687d16fe1789ba3aaa49f65fdfd66759a47f6769d92a146b9" gracePeriod=30 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.440953 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441063 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441100 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glwbn\" (UniqueName: \"kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441150 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441184 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfwf\" (UniqueName: \"kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441247 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441270 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441298 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441348 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441458 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.441768 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.442345 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.443761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.452913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.461258 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfwf\" (UniqueName: \"kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf\") pod \"horizon-776cbdfc85-bf5f7\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.542749 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.542811 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.542856 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glwbn\" (UniqueName: \"kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.542904 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.542922 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.543358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.543681 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.544372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.547689 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.560660 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glwbn\" (UniqueName: \"kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn\") pod \"horizon-68f4dc5599-pz5vk\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.562683 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.697801 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.707013 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a15d14e-bbad-4071-8531-ee330e283a2a" path="/var/lib/kubelet/pods/8a15d14e-bbad-4071-8531-ee330e283a2a/volumes" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.813945 4959 generic.go:334] "Generic (PLEG): container finished" podID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerID="b803fd51295e215eae9411738e10fe8923b5148585e6b75c37e71b66d6a252d4" exitCode=143 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.814020 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerDied","Data":"b803fd51295e215eae9411738e10fe8923b5148585e6b75c37e71b66d6a252d4"} Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.817092 4959 generic.go:334] "Generic (PLEG): container finished" podID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerID="b50cf27ea1530345a2dfa9e9f67ae26f85c66ad093aebeefc2e5dc15adb45c0c" exitCode=143 Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.817130 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerDied","Data":"b50cf27ea1530345a2dfa9e9f67ae26f85c66ad093aebeefc2e5dc15adb45c0c"} Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.841889 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.892026 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.894094 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:17 crc kubenswrapper[4959]: I1003 15:14:17.905482 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.045137 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.054278 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.054356 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.054433 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwftq\" (UniqueName: \"kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.054698 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.054768 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.156238 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwftq\" (UniqueName: \"kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.156479 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.156655 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.156840 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.157080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.157420 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.157781 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.159634 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.162299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.171485 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwftq\" (UniqueName: \"kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq\") pod \"horizon-cb4d9c449-s24wt\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.219004 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:18 crc kubenswrapper[4959]: W1003 15:14:18.220845 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7752fd42_9ac5_4ff8_8828_3bbf03b55f51.slice/crio-f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3 WatchSource:0}: Error finding container f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3: Status 404 returned error can't find the container with id f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3 Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.237217 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.758418 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.826570 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerStarted","Data":"f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3"} Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.828139 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerStarted","Data":"22d488197243bcacfaf50808d2b4ad524bbd8d1ae1edf133c2eb9056c569511e"} Oct 03 15:14:18 crc kubenswrapper[4959]: I1003 15:14:18.829538 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerStarted","Data":"b958d7f2d1dd0e6b49ed6fca9a0a067a3893feca955ed7ffc34c163722c5f81a"} Oct 03 15:14:20 crc kubenswrapper[4959]: I1003 15:14:20.853299 4959 generic.go:334] "Generic (PLEG): container finished" podID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerID="43f8f5ac309a5bb879d4e686502c58a47a358376d14d8d63685b476f5fa43540" exitCode=0 Oct 03 15:14:20 crc kubenswrapper[4959]: I1003 15:14:20.853342 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerDied","Data":"43f8f5ac309a5bb879d4e686502c58a47a358376d14d8d63685b476f5fa43540"} Oct 03 15:14:20 crc kubenswrapper[4959]: I1003 15:14:20.857353 4959 generic.go:334] "Generic (PLEG): container finished" podID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerID="2bbff60e17c67bb687d16fe1789ba3aaa49f65fdfd66759a47f6769d92a146b9" exitCode=0 Oct 03 15:14:20 crc kubenswrapper[4959]: I1003 15:14:20.857375 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerDied","Data":"2bbff60e17c67bb687d16fe1789ba3aaa49f65fdfd66759a47f6769d92a146b9"} Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.891360 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.896329 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.901744 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.967521 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.967912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7t8p\" (UniqueName: \"kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:22 crc kubenswrapper[4959]: I1003 15:14:22.968094 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.069700 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7t8p\" (UniqueName: \"kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.069792 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.069871 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.070488 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.070751 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.103401 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7t8p\" (UniqueName: \"kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p\") pod \"redhat-marketplace-c28kt\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:23 crc kubenswrapper[4959]: E1003 15:14:23.162276 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a72d18_2ef4_45d8_a807_eb7b3162bebf.slice/crio-94fd075459f614aa153845e7105d0cff40a675750d2119fb88ec1e4694626532\": RecentStats: unable to find data in memory cache]" Oct 03 15:14:23 crc kubenswrapper[4959]: I1003 15:14:23.228981 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.738622 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.809607 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9khr\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.809924 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.809989 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.810041 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.810200 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.810241 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.810273 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data\") pod \"d331bca5-1a85-4a21-9a10-912276df1ca3\" (UID: \"d331bca5-1a85-4a21-9a10-912276df1ca3\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.810647 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.811129 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs" (OuterVolumeSpecName: "logs") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.814862 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts" (OuterVolumeSpecName: "scripts") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.818539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph" (OuterVolumeSpecName: "ceph") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.823773 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr" (OuterVolumeSpecName: "kube-api-access-f9khr") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "kube-api-access-f9khr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.826127 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.877897 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.916110 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.916495 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.917219 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs" (OuterVolumeSpecName: "logs") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.917359 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.917386 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.918001 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.918369 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7flv\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.918441 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.918486 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle\") pod \"c546427c-b9ad-4355-867a-7b90a42bd5c9\" (UID: \"c546427c-b9ad-4355-867a-7b90a42bd5c9\") " Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920131 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920155 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920165 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920175 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920311 4959 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c546427c-b9ad-4355-867a-7b90a42bd5c9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920325 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d331bca5-1a85-4a21-9a10-912276df1ca3-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920358 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9khr\" (UniqueName: \"kubernetes.io/projected/d331bca5-1a85-4a21-9a10-912276df1ca3-kube-api-access-f9khr\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.920371 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.922505 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts" (OuterVolumeSpecName: "scripts") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.922989 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph" (OuterVolumeSpecName: "ceph") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.924279 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c546427c-b9ad-4355-867a-7b90a42bd5c9","Type":"ContainerDied","Data":"e02881ce0c0240589163b758f7727fce9b1c1195e47f4e223023444ad383be28"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.924321 4959 scope.go:117] "RemoveContainer" containerID="43f8f5ac309a5bb879d4e686502c58a47a358376d14d8d63685b476f5fa43540" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.924439 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.924749 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv" (OuterVolumeSpecName: "kube-api-access-x7flv") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "kube-api-access-x7flv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.926540 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerStarted","Data":"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.929571 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data" (OuterVolumeSpecName: "config-data") pod "d331bca5-1a85-4a21-9a10-912276df1ca3" (UID: "d331bca5-1a85-4a21-9a10-912276df1ca3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.944311 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.944824 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d331bca5-1a85-4a21-9a10-912276df1ca3","Type":"ContainerDied","Data":"f49facb801a5ac64a308caeaeaccc019ae8f51b268c1e034579edf7ee72352cb"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.948516 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.953251 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerStarted","Data":"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.953304 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerStarted","Data":"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.958717 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerStarted","Data":"9d87c1ccae0832808d108bad5ef0b40207f9632d771db885a7a866bb6cebec12"} Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.958872 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68f4dc5599-pz5vk" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon-log" containerID="cri-o://9d87c1ccae0832808d108bad5ef0b40207f9632d771db885a7a866bb6cebec12" gracePeriod=30 Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.958946 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68f4dc5599-pz5vk" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon" containerID="cri-o://6b12290d281b9f0de0443839cfc0ff13ded37637f2c656c14146753fe716bbab" gracePeriod=30 Oct 03 15:14:24 crc kubenswrapper[4959]: W1003 15:14:24.973628 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97298817_7fae_4b3b_8db4_493cb8b4c3bf.slice/crio-1403143303b930e2e85814c1be42d01f2341edf725772f2c1bcdff9c155df371 WatchSource:0}: Error finding container 1403143303b930e2e85814c1be42d01f2341edf725772f2c1bcdff9c155df371: Status 404 returned error can't find the container with id 1403143303b930e2e85814c1be42d01f2341edf725772f2c1bcdff9c155df371 Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.988815 4959 scope.go:117] "RemoveContainer" containerID="b803fd51295e215eae9411738e10fe8923b5148585e6b75c37e71b66d6a252d4" Oct 03 15:14:24 crc kubenswrapper[4959]: I1003 15:14:24.994182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.017553 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-776cbdfc85-bf5f7" podStartSLOduration=1.7350061879999998 podStartE2EDuration="8.017531623s" podCreationTimestamp="2025-10-03 15:14:17 +0000 UTC" firstStartedPulling="2025-10-03 15:14:18.058728478 +0000 UTC m=+6227.262071885" lastFinishedPulling="2025-10-03 15:14:24.341253903 +0000 UTC m=+6233.544597320" observedRunningTime="2025-10-03 15:14:24.980143589 +0000 UTC m=+6234.183487006" watchObservedRunningTime="2025-10-03 15:14:25.017531623 +0000 UTC m=+6234.220875040" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023562 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023597 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d331bca5-1a85-4a21-9a10-912276df1ca3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023608 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7flv\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-kube-api-access-x7flv\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023617 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023629 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c546427c-b9ad-4355-867a-7b90a42bd5c9-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.023742 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data" (OuterVolumeSpecName: "config-data") pod "c546427c-b9ad-4355-867a-7b90a42bd5c9" (UID: "c546427c-b9ad-4355-867a-7b90a42bd5c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.036183 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68f4dc5599-pz5vk" podStartSLOduration=1.9191746950000002 podStartE2EDuration="8.03615838s" podCreationTimestamp="2025-10-03 15:14:17 +0000 UTC" firstStartedPulling="2025-10-03 15:14:18.226017112 +0000 UTC m=+6227.429360529" lastFinishedPulling="2025-10-03 15:14:24.343000797 +0000 UTC m=+6233.546344214" observedRunningTime="2025-10-03 15:14:25.009978129 +0000 UTC m=+6234.213321546" watchObservedRunningTime="2025-10-03 15:14:25.03615838 +0000 UTC m=+6234.239501797" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.070989 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.081392 4959 scope.go:117] "RemoveContainer" containerID="2bbff60e17c67bb687d16fe1789ba3aaa49f65fdfd66759a47f6769d92a146b9" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.082716 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.090362 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-c8twv"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.099523 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-c8twv"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.110539 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: E1003 15:14:25.110962 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.110979 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: E1003 15:14:25.110994 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111001 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: E1003 15:14:25.111025 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111030 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: E1003 15:14:25.111055 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111060 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111278 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111293 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111309 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" containerName="glance-log" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.111319 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" containerName="glance-httpd" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.112365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.114659 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.115452 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.124659 4959 scope.go:117] "RemoveContainer" containerID="b50cf27ea1530345a2dfa9e9f67ae26f85c66ad093aebeefc2e5dc15adb45c0c" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.126094 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c546427c-b9ad-4355-867a-7b90a42bd5c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.227921 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229249 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229283 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grhj8\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-kube-api-access-grhj8\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229352 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229377 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229414 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.229431 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.328554 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331385 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331420 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331459 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331562 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.331592 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grhj8\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-kube-api-access-grhj8\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.332038 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-logs\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.332100 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.336255 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.336507 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.337046 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.337903 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.340526 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.371559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grhj8\" (UniqueName: \"kubernetes.io/projected/c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03-kube-api-access-grhj8\") pod \"glance-default-internal-api-0\" (UID: \"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03\") " pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.385028 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.386715 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.389524 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.416912 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.443502 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536627 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56s6v\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-kube-api-access-56s6v\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536667 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536751 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-logs\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536814 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-ceph\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536903 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536968 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.536993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.638913 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56s6v\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-kube-api-access-56s6v\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639248 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639283 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-logs\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639758 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-ceph\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639785 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-logs\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.639938 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.640027 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.640069 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.644438 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.644860 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-scripts\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.644913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-config-data\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.645298 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-ceph\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.657077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56s6v\" (UniqueName: \"kubernetes.io/projected/9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7-kube-api-access-56s6v\") pod \"glance-default-external-api-0\" (UID: \"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7\") " pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.701128 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949ca17b-cb34-493c-b135-300fb15c2e1a" path="/var/lib/kubelet/pods/949ca17b-cb34-493c-b135-300fb15c2e1a/volumes" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.701759 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c546427c-b9ad-4355-867a-7b90a42bd5c9" path="/var/lib/kubelet/pods/c546427c-b9ad-4355-867a-7b90a42bd5c9/volumes" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.702552 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d331bca5-1a85-4a21-9a10-912276df1ca3" path="/var/lib/kubelet/pods/d331bca5-1a85-4a21-9a10-912276df1ca3/volumes" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.746747 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.979268 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerStarted","Data":"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0"} Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.984016 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.984639 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerStarted","Data":"6b12290d281b9f0de0443839cfc0ff13ded37637f2c656c14146753fe716bbab"} Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.987098 4959 generic.go:334] "Generic (PLEG): container finished" podID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerID="e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53" exitCode=0 Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.987161 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerDied","Data":"e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53"} Oct 03 15:14:25 crc kubenswrapper[4959]: I1003 15:14:25.987226 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerStarted","Data":"1403143303b930e2e85814c1be42d01f2341edf725772f2c1bcdff9c155df371"} Oct 03 15:14:25 crc kubenswrapper[4959]: W1003 15:14:25.994139 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5185f5a_c1ec_4beb_9b1d_79f01f3a0d03.slice/crio-d406dea8930ee3f5c58ce96043dd950ee6abd638423de5dda0a37e9f46b4df56 WatchSource:0}: Error finding container d406dea8930ee3f5c58ce96043dd950ee6abd638423de5dda0a37e9f46b4df56: Status 404 returned error can't find the container with id d406dea8930ee3f5c58ce96043dd950ee6abd638423de5dda0a37e9f46b4df56 Oct 03 15:14:26 crc kubenswrapper[4959]: I1003 15:14:26.021456 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-cb4d9c449-s24wt" podStartSLOduration=3.442693449 podStartE2EDuration="9.021438491s" podCreationTimestamp="2025-10-03 15:14:17 +0000 UTC" firstStartedPulling="2025-10-03 15:14:18.764825398 +0000 UTC m=+6227.968168815" lastFinishedPulling="2025-10-03 15:14:24.34357044 +0000 UTC m=+6233.546913857" observedRunningTime="2025-10-03 15:14:26.009462638 +0000 UTC m=+6235.212806075" watchObservedRunningTime="2025-10-03 15:14:26.021438491 +0000 UTC m=+6235.224781908" Oct 03 15:14:26 crc kubenswrapper[4959]: I1003 15:14:26.300976 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.009591 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7","Type":"ContainerStarted","Data":"ff1e7ce0e7de3ea0c4eb6e772963b31338cf16c20687ea692a8a8858ee5c6633"} Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.010229 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7","Type":"ContainerStarted","Data":"195b82d1271296de2f924f8bb53041242c56a4eff106dfb7d82cb7d508f0ab1f"} Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.015505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03","Type":"ContainerStarted","Data":"1a648978062e362dd2b63dd3deb128d9dd347a5e381df8a71e112460355c3727"} Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.015548 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03","Type":"ContainerStarted","Data":"d406dea8930ee3f5c58ce96043dd950ee6abd638423de5dda0a37e9f46b4df56"} Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.562829 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.563125 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:27 crc kubenswrapper[4959]: I1003 15:14:27.705683 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.033333 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03","Type":"ContainerStarted","Data":"f17cd19fff418b00024f58cc2663907cf15085900ebd2839a3bd943d515f56ad"} Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.040287 4959 generic.go:334] "Generic (PLEG): container finished" podID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerID="24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a" exitCode=0 Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.040372 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerDied","Data":"24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a"} Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.046062 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7","Type":"ContainerStarted","Data":"40dc37fb7f63391cf15ec056348f81003f908dbf4a3e8dbef92d489509307c91"} Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.062036 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.062013258 podStartE2EDuration="3.062013258s" podCreationTimestamp="2025-10-03 15:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:14:28.054270119 +0000 UTC m=+6237.257613576" watchObservedRunningTime="2025-10-03 15:14:28.062013258 +0000 UTC m=+6237.265356675" Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.112701 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.112679238 podStartE2EDuration="3.112679238s" podCreationTimestamp="2025-10-03 15:14:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:14:28.097869665 +0000 UTC m=+6237.301213092" watchObservedRunningTime="2025-10-03 15:14:28.112679238 +0000 UTC m=+6237.316022675" Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.237892 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:28 crc kubenswrapper[4959]: I1003 15:14:28.237954 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:30 crc kubenswrapper[4959]: I1003 15:14:30.081732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerStarted","Data":"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9"} Oct 03 15:14:30 crc kubenswrapper[4959]: I1003 15:14:30.102358 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c28kt" podStartSLOduration=4.899067118 podStartE2EDuration="8.102332059s" podCreationTimestamp="2025-10-03 15:14:22 +0000 UTC" firstStartedPulling="2025-10-03 15:14:25.989465438 +0000 UTC m=+6235.192808855" lastFinishedPulling="2025-10-03 15:14:29.192730369 +0000 UTC m=+6238.396073796" observedRunningTime="2025-10-03 15:14:30.096763113 +0000 UTC m=+6239.300106530" watchObservedRunningTime="2025-10-03 15:14:30.102332059 +0000 UTC m=+6239.305675476" Oct 03 15:14:31 crc kubenswrapper[4959]: E1003 15:14:31.764336 4959 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/34891f6711c829858e1b4eca44725f0e00fb7bd5c538a50fb181ec524e9f8ba6/diff" to get inode usage: stat /var/lib/containers/storage/overlay/34891f6711c829858e1b4eca44725f0e00fb7bd5c538a50fb181ec524e9f8ba6/diff: no such file or directory, extraDiskErr: Oct 03 15:14:33 crc kubenswrapper[4959]: I1003 15:14:33.230108 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:33 crc kubenswrapper[4959]: I1003 15:14:33.230438 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:33 crc kubenswrapper[4959]: I1003 15:14:33.286681 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:34 crc kubenswrapper[4959]: I1003 15:14:34.178092 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:34 crc kubenswrapper[4959]: I1003 15:14:34.224447 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.444303 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.444585 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.487517 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.496765 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.747247 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.748452 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.808244 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 15:14:35 crc kubenswrapper[4959]: I1003 15:14:35.816056 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.149055 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.149363 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.149030 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c28kt" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="registry-server" containerID="cri-o://041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9" gracePeriod=2 Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.149686 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.149707 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.788849 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.914338 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content\") pod \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.914454 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities\") pod \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.914490 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7t8p\" (UniqueName: \"kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p\") pod \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\" (UID: \"97298817-7fae-4b3b-8db4-493cb8b4c3bf\") " Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.916689 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities" (OuterVolumeSpecName: "utilities") pod "97298817-7fae-4b3b-8db4-493cb8b4c3bf" (UID: "97298817-7fae-4b3b-8db4-493cb8b4c3bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.922360 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p" (OuterVolumeSpecName: "kube-api-access-c7t8p") pod "97298817-7fae-4b3b-8db4-493cb8b4c3bf" (UID: "97298817-7fae-4b3b-8db4-493cb8b4c3bf"). InnerVolumeSpecName "kube-api-access-c7t8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:36 crc kubenswrapper[4959]: I1003 15:14:36.933678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97298817-7fae-4b3b-8db4-493cb8b4c3bf" (UID: "97298817-7fae-4b3b-8db4-493cb8b4c3bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.016999 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.017046 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97298817-7fae-4b3b-8db4-493cb8b4c3bf-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.017061 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7t8p\" (UniqueName: \"kubernetes.io/projected/97298817-7fae-4b3b-8db4-493cb8b4c3bf-kube-api-access-c7t8p\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.166314 4959 generic.go:334] "Generic (PLEG): container finished" podID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerID="041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9" exitCode=0 Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.166384 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c28kt" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.166452 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerDied","Data":"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9"} Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.166526 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c28kt" event={"ID":"97298817-7fae-4b3b-8db4-493cb8b4c3bf","Type":"ContainerDied","Data":"1403143303b930e2e85814c1be42d01f2341edf725772f2c1bcdff9c155df371"} Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.166557 4959 scope.go:117] "RemoveContainer" containerID="041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.224790 4959 scope.go:117] "RemoveContainer" containerID="24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.228630 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.242690 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c28kt"] Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.249391 4959 scope.go:117] "RemoveContainer" containerID="e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.309884 4959 scope.go:117] "RemoveContainer" containerID="041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9" Oct 03 15:14:37 crc kubenswrapper[4959]: E1003 15:14:37.310430 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9\": container with ID starting with 041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9 not found: ID does not exist" containerID="041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.310469 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9"} err="failed to get container status \"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9\": rpc error: code = NotFound desc = could not find container \"041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9\": container with ID starting with 041250f1b6b9e01a2c1b063657c4a649cc5f0fbfcfca8b7faa43d57c8961f6e9 not found: ID does not exist" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.310496 4959 scope.go:117] "RemoveContainer" containerID="24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a" Oct 03 15:14:37 crc kubenswrapper[4959]: E1003 15:14:37.311639 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a\": container with ID starting with 24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a not found: ID does not exist" containerID="24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.311679 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a"} err="failed to get container status \"24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a\": rpc error: code = NotFound desc = could not find container \"24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a\": container with ID starting with 24c6b84e7d3389ea0b1a7738bef23c6c54105ce900a861d311b9adf34a75707a not found: ID does not exist" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.311706 4959 scope.go:117] "RemoveContainer" containerID="e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53" Oct 03 15:14:37 crc kubenswrapper[4959]: E1003 15:14:37.311987 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53\": container with ID starting with e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53 not found: ID does not exist" containerID="e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.312013 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53"} err="failed to get container status \"e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53\": rpc error: code = NotFound desc = could not find container \"e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53\": container with ID starting with e06368d1291d1f5a87511a4535e333f8afbc70ca5e98d9c5ac59a259971b2d53 not found: ID does not exist" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.564142 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 03 15:14:37 crc kubenswrapper[4959]: I1003 15:14:37.698459 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" path="/var/lib/kubelet/pods/97298817-7fae-4b3b-8db4-493cb8b4c3bf/volumes" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.239751 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.286679 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.286794 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.560575 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.591963 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.592074 4959 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 15:14:38 crc kubenswrapper[4959]: I1003 15:14:38.593386 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 15:14:49 crc kubenswrapper[4959]: I1003 15:14:49.338002 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:50 crc kubenswrapper[4959]: I1003 15:14:50.087536 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:50 crc kubenswrapper[4959]: I1003 15:14:50.870412 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:14:51 crc kubenswrapper[4959]: I1003 15:14:51.811262 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:14:51 crc kubenswrapper[4959]: I1003 15:14:51.896815 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:14:51 crc kubenswrapper[4959]: I1003 15:14:51.897233 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon-log" containerID="cri-o://a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667" gracePeriod=30 Oct 03 15:14:51 crc kubenswrapper[4959]: I1003 15:14:51.897660 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" containerID="cri-o://869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42" gracePeriod=30 Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.349685 4959 generic.go:334] "Generic (PLEG): container finished" podID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerID="869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42" exitCode=0 Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.349732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerDied","Data":"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42"} Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353781 4959 generic.go:334] "Generic (PLEG): container finished" podID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerID="6b12290d281b9f0de0443839cfc0ff13ded37637f2c656c14146753fe716bbab" exitCode=137 Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353815 4959 generic.go:334] "Generic (PLEG): container finished" podID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerID="9d87c1ccae0832808d108bad5ef0b40207f9632d771db885a7a866bb6cebec12" exitCode=137 Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353839 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerDied","Data":"6b12290d281b9f0de0443839cfc0ff13ded37637f2c656c14146753fe716bbab"} Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353865 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerDied","Data":"9d87c1ccae0832808d108bad5ef0b40207f9632d771db885a7a866bb6cebec12"} Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353879 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68f4dc5599-pz5vk" event={"ID":"7752fd42-9ac5-4ff8-8828-3bbf03b55f51","Type":"ContainerDied","Data":"f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3"} Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.353891 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4929f691f424e1104e7c5ae9bc933f18e9db29962ee04352c071219e96271f3" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.399715 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.509731 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs\") pod \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.509953 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data\") pod \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.509977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glwbn\" (UniqueName: \"kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn\") pod \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.510105 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key\") pod \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.510160 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts\") pod \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\" (UID: \"7752fd42-9ac5-4ff8-8828-3bbf03b55f51\") " Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.510493 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs" (OuterVolumeSpecName: "logs") pod "7752fd42-9ac5-4ff8-8828-3bbf03b55f51" (UID: "7752fd42-9ac5-4ff8-8828-3bbf03b55f51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.518937 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn" (OuterVolumeSpecName: "kube-api-access-glwbn") pod "7752fd42-9ac5-4ff8-8828-3bbf03b55f51" (UID: "7752fd42-9ac5-4ff8-8828-3bbf03b55f51"). InnerVolumeSpecName "kube-api-access-glwbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.520132 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7752fd42-9ac5-4ff8-8828-3bbf03b55f51" (UID: "7752fd42-9ac5-4ff8-8828-3bbf03b55f51"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.548285 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts" (OuterVolumeSpecName: "scripts") pod "7752fd42-9ac5-4ff8-8828-3bbf03b55f51" (UID: "7752fd42-9ac5-4ff8-8828-3bbf03b55f51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.548740 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data" (OuterVolumeSpecName: "config-data") pod "7752fd42-9ac5-4ff8-8828-3bbf03b55f51" (UID: "7752fd42-9ac5-4ff8-8828-3bbf03b55f51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.612736 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.612772 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.612781 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.612792 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:55 crc kubenswrapper[4959]: I1003 15:14:55.612800 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glwbn\" (UniqueName: \"kubernetes.io/projected/7752fd42-9ac5-4ff8-8828-3bbf03b55f51-kube-api-access-glwbn\") on node \"crc\" DevicePath \"\"" Oct 03 15:14:56 crc kubenswrapper[4959]: I1003 15:14:56.362747 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68f4dc5599-pz5vk" Oct 03 15:14:56 crc kubenswrapper[4959]: I1003 15:14:56.385550 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:56 crc kubenswrapper[4959]: I1003 15:14:56.393159 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68f4dc5599-pz5vk"] Oct 03 15:14:57 crc kubenswrapper[4959]: I1003 15:14:57.563675 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 03 15:14:57 crc kubenswrapper[4959]: I1003 15:14:57.712124 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" path="/var/lib/kubelet/pods/7752fd42-9ac5-4ff8-8828-3bbf03b55f51/volumes" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.166654 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8"] Oct 03 15:15:00 crc kubenswrapper[4959]: E1003 15:15:00.168437 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.168477 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon" Oct 03 15:15:00 crc kubenswrapper[4959]: E1003 15:15:00.168508 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="extract-utilities" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.168523 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="extract-utilities" Oct 03 15:15:00 crc kubenswrapper[4959]: E1003 15:15:00.168558 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon-log" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.168575 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon-log" Oct 03 15:15:00 crc kubenswrapper[4959]: E1003 15:15:00.168608 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="registry-server" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.168621 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="registry-server" Oct 03 15:15:00 crc kubenswrapper[4959]: E1003 15:15:00.168674 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="extract-content" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.168688 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="extract-content" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.169073 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon-log" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.169119 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7752fd42-9ac5-4ff8-8828-3bbf03b55f51" containerName="horizon" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.169158 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="97298817-7fae-4b3b-8db4-493cb8b4c3bf" containerName="registry-server" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.170743 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.173998 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.174084 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.179911 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8"] Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.247466 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xph9s\" (UniqueName: \"kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.248026 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.248071 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.350682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xph9s\" (UniqueName: \"kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.351050 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.351178 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.352437 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.361676 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.375248 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xph9s\" (UniqueName: \"kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s\") pod \"collect-profiles-29325075-29nr8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:00 crc kubenswrapper[4959]: I1003 15:15:00.510828 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:01 crc kubenswrapper[4959]: I1003 15:15:01.016921 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8"] Oct 03 15:15:01 crc kubenswrapper[4959]: I1003 15:15:01.417462 4959 generic.go:334] "Generic (PLEG): container finished" podID="1833901d-4e6a-495f-a1d4-e6b0221acdd8" containerID="dc269c41f329990be68dd79a167c80089053f01589762f1624ac83de8de37b29" exitCode=0 Oct 03 15:15:01 crc kubenswrapper[4959]: I1003 15:15:01.417587 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" event={"ID":"1833901d-4e6a-495f-a1d4-e6b0221acdd8","Type":"ContainerDied","Data":"dc269c41f329990be68dd79a167c80089053f01589762f1624ac83de8de37b29"} Oct 03 15:15:01 crc kubenswrapper[4959]: I1003 15:15:01.417798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" event={"ID":"1833901d-4e6a-495f-a1d4-e6b0221acdd8","Type":"ContainerStarted","Data":"cd413dc968ad20164717032cdc6a4f5d008b1b069652b4080508e98e5168d148"} Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.800689 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.910271 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume\") pod \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.910370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume\") pod \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.910647 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xph9s\" (UniqueName: \"kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s\") pod \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\" (UID: \"1833901d-4e6a-495f-a1d4-e6b0221acdd8\") " Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.911102 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume" (OuterVolumeSpecName: "config-volume") pod "1833901d-4e6a-495f-a1d4-e6b0221acdd8" (UID: "1833901d-4e6a-495f-a1d4-e6b0221acdd8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.911811 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1833901d-4e6a-495f-a1d4-e6b0221acdd8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.916805 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1833901d-4e6a-495f-a1d4-e6b0221acdd8" (UID: "1833901d-4e6a-495f-a1d4-e6b0221acdd8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:15:02 crc kubenswrapper[4959]: I1003 15:15:02.917771 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s" (OuterVolumeSpecName: "kube-api-access-xph9s") pod "1833901d-4e6a-495f-a1d4-e6b0221acdd8" (UID: "1833901d-4e6a-495f-a1d4-e6b0221acdd8"). InnerVolumeSpecName "kube-api-access-xph9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.015426 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xph9s\" (UniqueName: \"kubernetes.io/projected/1833901d-4e6a-495f-a1d4-e6b0221acdd8-kube-api-access-xph9s\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.015486 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1833901d-4e6a-495f-a1d4-e6b0221acdd8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.446154 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" event={"ID":"1833901d-4e6a-495f-a1d4-e6b0221acdd8","Type":"ContainerDied","Data":"cd413dc968ad20164717032cdc6a4f5d008b1b069652b4080508e98e5168d148"} Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.446227 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd413dc968ad20164717032cdc6a4f5d008b1b069652b4080508e98e5168d148" Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.446290 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8" Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.889780 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt"] Oct 03 15:15:03 crc kubenswrapper[4959]: I1003 15:15:03.902921 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325030-pt8dt"] Oct 03 15:15:05 crc kubenswrapper[4959]: I1003 15:15:05.700018 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb063e10-6487-48b0-84c9-258187deb68b" path="/var/lib/kubelet/pods/cb063e10-6487-48b0-84c9-258187deb68b/volumes" Oct 03 15:15:07 crc kubenswrapper[4959]: I1003 15:15:07.165888 4959 scope.go:117] "RemoveContainer" containerID="d7a3dca16749d3380a201ff1a00c43a84b63e047384436423e1c4c1921fed001" Oct 03 15:15:07 crc kubenswrapper[4959]: I1003 15:15:07.219662 4959 scope.go:117] "RemoveContainer" containerID="097bf3638a3d687477aea94d3077f7ed9a8248820dbb5b43bd3d78a34fe62d3b" Oct 03 15:15:07 crc kubenswrapper[4959]: I1003 15:15:07.244638 4959 scope.go:117] "RemoveContainer" containerID="e43673f3a56fb8a117fe1293e702428f6566e2d8f81258b79981b349fcd1d50f" Oct 03 15:15:07 crc kubenswrapper[4959]: I1003 15:15:07.284131 4959 scope.go:117] "RemoveContainer" containerID="f17fdd0fa725c3ba22091e35e6dd8f60c47ccebcabf002a913e97455899d93ce" Oct 03 15:15:07 crc kubenswrapper[4959]: I1003 15:15:07.564261 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 03 15:15:08 crc kubenswrapper[4959]: I1003 15:15:08.046702 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-mndcs"] Oct 03 15:15:08 crc kubenswrapper[4959]: I1003 15:15:08.057386 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-mndcs"] Oct 03 15:15:09 crc kubenswrapper[4959]: I1003 15:15:09.702079 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c" path="/var/lib/kubelet/pods/9ea6cb3d-03f4-4eb2-8d94-ac9a9a8cf37c/volumes" Oct 03 15:15:17 crc kubenswrapper[4959]: I1003 15:15:17.046225 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-10f6-account-create-hmtcb"] Oct 03 15:15:17 crc kubenswrapper[4959]: I1003 15:15:17.059689 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-10f6-account-create-hmtcb"] Oct 03 15:15:17 crc kubenswrapper[4959]: I1003 15:15:17.563401 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776cbdfc85-bf5f7" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 03 15:15:17 crc kubenswrapper[4959]: I1003 15:15:17.563778 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:15:17 crc kubenswrapper[4959]: I1003 15:15:17.697012 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f542467-086c-4c9d-b634-d94240cdc81a" path="/var/lib/kubelet/pods/2f542467-086c-4c9d-b634-d94240cdc81a/volumes" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.289215 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.350516 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts\") pod \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.350587 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfwf\" (UniqueName: \"kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf\") pod \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.350634 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs\") pod \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.350668 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data\") pod \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.350754 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key\") pod \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\" (UID: \"4eb1516b-193a-42dc-8104-aa3df0e2ed60\") " Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.351730 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs" (OuterVolumeSpecName: "logs") pod "4eb1516b-193a-42dc-8104-aa3df0e2ed60" (UID: "4eb1516b-193a-42dc-8104-aa3df0e2ed60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.356695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4eb1516b-193a-42dc-8104-aa3df0e2ed60" (UID: "4eb1516b-193a-42dc-8104-aa3df0e2ed60"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.358781 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf" (OuterVolumeSpecName: "kube-api-access-8dfwf") pod "4eb1516b-193a-42dc-8104-aa3df0e2ed60" (UID: "4eb1516b-193a-42dc-8104-aa3df0e2ed60"). InnerVolumeSpecName "kube-api-access-8dfwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.376541 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data" (OuterVolumeSpecName: "config-data") pod "4eb1516b-193a-42dc-8104-aa3df0e2ed60" (UID: "4eb1516b-193a-42dc-8104-aa3df0e2ed60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.379578 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts" (OuterVolumeSpecName: "scripts") pod "4eb1516b-193a-42dc-8104-aa3df0e2ed60" (UID: "4eb1516b-193a-42dc-8104-aa3df0e2ed60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.453482 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.453535 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfwf\" (UniqueName: \"kubernetes.io/projected/4eb1516b-193a-42dc-8104-aa3df0e2ed60-kube-api-access-8dfwf\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.453549 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4eb1516b-193a-42dc-8104-aa3df0e2ed60-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.453558 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4eb1516b-193a-42dc-8104-aa3df0e2ed60-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.453569 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4eb1516b-193a-42dc-8104-aa3df0e2ed60-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.698840 4959 generic.go:334] "Generic (PLEG): container finished" podID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerID="a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667" exitCode=137 Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.698905 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerDied","Data":"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667"} Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.698979 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776cbdfc85-bf5f7" event={"ID":"4eb1516b-193a-42dc-8104-aa3df0e2ed60","Type":"ContainerDied","Data":"b958d7f2d1dd0e6b49ed6fca9a0a067a3893feca955ed7ffc34c163722c5f81a"} Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.699006 4959 scope.go:117] "RemoveContainer" containerID="869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.699272 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776cbdfc85-bf5f7" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.739842 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.746833 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-776cbdfc85-bf5f7"] Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.899166 4959 scope.go:117] "RemoveContainer" containerID="a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.920042 4959 scope.go:117] "RemoveContainer" containerID="869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42" Oct 03 15:15:22 crc kubenswrapper[4959]: E1003 15:15:22.920771 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42\": container with ID starting with 869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42 not found: ID does not exist" containerID="869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.920797 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42"} err="failed to get container status \"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42\": rpc error: code = NotFound desc = could not find container \"869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42\": container with ID starting with 869d7a8798ac84f4954e0fac904d7a14412dd60d3ce0901141e371172f72bf42 not found: ID does not exist" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.920819 4959 scope.go:117] "RemoveContainer" containerID="a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667" Oct 03 15:15:22 crc kubenswrapper[4959]: E1003 15:15:22.921035 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667\": container with ID starting with a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667 not found: ID does not exist" containerID="a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667" Oct 03 15:15:22 crc kubenswrapper[4959]: I1003 15:15:22.921068 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667"} err="failed to get container status \"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667\": rpc error: code = NotFound desc = could not find container \"a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667\": container with ID starting with a51454a5063a59cbc5601e7b2f6b858c2e7e4f26ca57bcf9d428f391ea636667 not found: ID does not exist" Oct 03 15:15:23 crc kubenswrapper[4959]: I1003 15:15:23.700001 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" path="/var/lib/kubelet/pods/4eb1516b-193a-42dc-8104-aa3df0e2ed60/volumes" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.648881 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5656c49bff-fgjq5"] Oct 03 15:15:25 crc kubenswrapper[4959]: E1003 15:15:25.650485 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1833901d-4e6a-495f-a1d4-e6b0221acdd8" containerName="collect-profiles" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.650599 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1833901d-4e6a-495f-a1d4-e6b0221acdd8" containerName="collect-profiles" Oct 03 15:15:25 crc kubenswrapper[4959]: E1003 15:15:25.650678 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon-log" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.650733 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon-log" Oct 03 15:15:25 crc kubenswrapper[4959]: E1003 15:15:25.650809 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.650862 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.651088 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1833901d-4e6a-495f-a1d4-e6b0221acdd8" containerName="collect-profiles" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.651154 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.651262 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb1516b-193a-42dc-8104-aa3df0e2ed60" containerName="horizon-log" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.655450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.667497 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5656c49bff-fgjq5"] Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.721123 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-scripts\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.721413 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ce9411f-dcdd-4538-a83e-193382cda8e3-horizon-secret-key\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.721513 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwxwx\" (UniqueName: \"kubernetes.io/projected/8ce9411f-dcdd-4538-a83e-193382cda8e3-kube-api-access-zwxwx\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.721618 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-config-data\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.722424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce9411f-dcdd-4538-a83e-193382cda8e3-logs\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.823589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-config-data\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.823629 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce9411f-dcdd-4538-a83e-193382cda8e3-logs\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.823729 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-scripts\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.823767 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ce9411f-dcdd-4538-a83e-193382cda8e3-horizon-secret-key\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.823798 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwxwx\" (UniqueName: \"kubernetes.io/projected/8ce9411f-dcdd-4538-a83e-193382cda8e3-kube-api-access-zwxwx\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.825139 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-config-data\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.825386 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce9411f-dcdd-4538-a83e-193382cda8e3-logs\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.825803 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8ce9411f-dcdd-4538-a83e-193382cda8e3-scripts\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.838107 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8ce9411f-dcdd-4538-a83e-193382cda8e3-horizon-secret-key\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.843171 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwxwx\" (UniqueName: \"kubernetes.io/projected/8ce9411f-dcdd-4538-a83e-193382cda8e3-kube-api-access-zwxwx\") pod \"horizon-5656c49bff-fgjq5\" (UID: \"8ce9411f-dcdd-4538-a83e-193382cda8e3\") " pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:25 crc kubenswrapper[4959]: I1003 15:15:25.979852 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.056162 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-klb8j"] Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.063694 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-klb8j"] Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.487405 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5656c49bff-fgjq5"] Oct 03 15:15:26 crc kubenswrapper[4959]: W1003 15:15:26.509425 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ce9411f_dcdd_4538_a83e_193382cda8e3.slice/crio-0955d8ba23e16d9b68c6ca1739a007ca828067befc3aa8f1ebb88945e086d2db WatchSource:0}: Error finding container 0955d8ba23e16d9b68c6ca1739a007ca828067befc3aa8f1ebb88945e086d2db: Status 404 returned error can't find the container with id 0955d8ba23e16d9b68c6ca1739a007ca828067befc3aa8f1ebb88945e086d2db Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.735362 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-sbh94"] Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.737831 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-sbh94" Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.738857 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5656c49bff-fgjq5" event={"ID":"8ce9411f-dcdd-4538-a83e-193382cda8e3","Type":"ContainerStarted","Data":"36d8844a712350a69a19ec9cf8f8f4fc483b8f90304d3ecf730b231a471073af"} Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.738898 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5656c49bff-fgjq5" event={"ID":"8ce9411f-dcdd-4538-a83e-193382cda8e3","Type":"ContainerStarted","Data":"0955d8ba23e16d9b68c6ca1739a007ca828067befc3aa8f1ebb88945e086d2db"} Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.740283 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcvf5\" (UniqueName: \"kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5\") pod \"heat-db-create-sbh94\" (UID: \"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689\") " pod="openstack/heat-db-create-sbh94" Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.754105 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-sbh94"] Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.843328 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcvf5\" (UniqueName: \"kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5\") pod \"heat-db-create-sbh94\" (UID: \"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689\") " pod="openstack/heat-db-create-sbh94" Oct 03 15:15:26 crc kubenswrapper[4959]: I1003 15:15:26.861602 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcvf5\" (UniqueName: \"kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5\") pod \"heat-db-create-sbh94\" (UID: \"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689\") " pod="openstack/heat-db-create-sbh94" Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.057103 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-sbh94" Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.538610 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-sbh94"] Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.698049 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2c7fceb-49e7-47ca-940b-d2fc816c4f8f" path="/var/lib/kubelet/pods/b2c7fceb-49e7-47ca-940b-d2fc816c4f8f/volumes" Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.762245 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-sbh94" event={"ID":"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689","Type":"ContainerStarted","Data":"de08e4d4afcd1d13c69f2957aa766f09fa17fca5a655c50601d1d11b54cefed4"} Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.763913 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-sbh94" event={"ID":"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689","Type":"ContainerStarted","Data":"cb5097a497feff4644f83371571b364506215c97ebe77b01dce0635f4d298e5f"} Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.765633 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5656c49bff-fgjq5" event={"ID":"8ce9411f-dcdd-4538-a83e-193382cda8e3","Type":"ContainerStarted","Data":"ad4c0bb75eafdc581b756d5ac48d6d158b4a01d7ea32edd5062b3d5a289dd2c3"} Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.777466 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-sbh94" podStartSLOduration=1.777450446 podStartE2EDuration="1.777450446s" podCreationTimestamp="2025-10-03 15:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:15:27.776312518 +0000 UTC m=+6296.979655935" watchObservedRunningTime="2025-10-03 15:15:27.777450446 +0000 UTC m=+6296.980793863" Oct 03 15:15:27 crc kubenswrapper[4959]: I1003 15:15:27.806497 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5656c49bff-fgjq5" podStartSLOduration=2.806474276 podStartE2EDuration="2.806474276s" podCreationTimestamp="2025-10-03 15:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:15:27.796975984 +0000 UTC m=+6297.000319441" watchObservedRunningTime="2025-10-03 15:15:27.806474276 +0000 UTC m=+6297.009817693" Oct 03 15:15:28 crc kubenswrapper[4959]: I1003 15:15:28.785986 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-sbh94" event={"ID":"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689","Type":"ContainerDied","Data":"de08e4d4afcd1d13c69f2957aa766f09fa17fca5a655c50601d1d11b54cefed4"} Oct 03 15:15:28 crc kubenswrapper[4959]: I1003 15:15:28.784867 4959 generic.go:334] "Generic (PLEG): container finished" podID="e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" containerID="de08e4d4afcd1d13c69f2957aa766f09fa17fca5a655c50601d1d11b54cefed4" exitCode=0 Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.216358 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-sbh94" Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.223925 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcvf5\" (UniqueName: \"kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5\") pod \"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689\" (UID: \"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689\") " Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.230136 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5" (OuterVolumeSpecName: "kube-api-access-tcvf5") pod "e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" (UID: "e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689"). InnerVolumeSpecName "kube-api-access-tcvf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.326628 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcvf5\" (UniqueName: \"kubernetes.io/projected/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689-kube-api-access-tcvf5\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.814872 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-sbh94" event={"ID":"e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689","Type":"ContainerDied","Data":"cb5097a497feff4644f83371571b364506215c97ebe77b01dce0635f4d298e5f"} Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.814911 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-sbh94" Oct 03 15:15:30 crc kubenswrapper[4959]: I1003 15:15:30.814921 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb5097a497feff4644f83371571b364506215c97ebe77b01dce0635f4d298e5f" Oct 03 15:15:35 crc kubenswrapper[4959]: I1003 15:15:35.980897 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:35 crc kubenswrapper[4959]: I1003 15:15:35.981464 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.045126 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.045233 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.981715 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-d30d-account-create-cfnvj"] Oct 03 15:15:36 crc kubenswrapper[4959]: E1003 15:15:36.982447 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" containerName="mariadb-database-create" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.982465 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" containerName="mariadb-database-create" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.982695 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" containerName="mariadb-database-create" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.983391 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:36 crc kubenswrapper[4959]: I1003 15:15:36.985479 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 03 15:15:37 crc kubenswrapper[4959]: I1003 15:15:37.004662 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d30d-account-create-cfnvj"] Oct 03 15:15:37 crc kubenswrapper[4959]: I1003 15:15:37.174747 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67qzk\" (UniqueName: \"kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk\") pod \"heat-d30d-account-create-cfnvj\" (UID: \"ebce59c6-74ef-431c-8685-ee7eaf0c319a\") " pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:37 crc kubenswrapper[4959]: I1003 15:15:37.276656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67qzk\" (UniqueName: \"kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk\") pod \"heat-d30d-account-create-cfnvj\" (UID: \"ebce59c6-74ef-431c-8685-ee7eaf0c319a\") " pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:37 crc kubenswrapper[4959]: I1003 15:15:37.317157 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67qzk\" (UniqueName: \"kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk\") pod \"heat-d30d-account-create-cfnvj\" (UID: \"ebce59c6-74ef-431c-8685-ee7eaf0c319a\") " pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:37 crc kubenswrapper[4959]: I1003 15:15:37.604313 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:38 crc kubenswrapper[4959]: I1003 15:15:38.078474 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d30d-account-create-cfnvj"] Oct 03 15:15:38 crc kubenswrapper[4959]: I1003 15:15:38.903264 4959 generic.go:334] "Generic (PLEG): container finished" podID="ebce59c6-74ef-431c-8685-ee7eaf0c319a" containerID="041b30bde2e90f807045461e862856e8b2143bdf81845e0ab042fe9b8d90f473" exitCode=0 Oct 03 15:15:38 crc kubenswrapper[4959]: I1003 15:15:38.903398 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d30d-account-create-cfnvj" event={"ID":"ebce59c6-74ef-431c-8685-ee7eaf0c319a","Type":"ContainerDied","Data":"041b30bde2e90f807045461e862856e8b2143bdf81845e0ab042fe9b8d90f473"} Oct 03 15:15:38 crc kubenswrapper[4959]: I1003 15:15:38.903685 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d30d-account-create-cfnvj" event={"ID":"ebce59c6-74ef-431c-8685-ee7eaf0c319a","Type":"ContainerStarted","Data":"f153f06182686ba4fcdab31b10990cac7ca989a848c0e15d3005fed432e4a8e1"} Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.400973 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.507756 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67qzk\" (UniqueName: \"kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk\") pod \"ebce59c6-74ef-431c-8685-ee7eaf0c319a\" (UID: \"ebce59c6-74ef-431c-8685-ee7eaf0c319a\") " Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.517504 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk" (OuterVolumeSpecName: "kube-api-access-67qzk") pod "ebce59c6-74ef-431c-8685-ee7eaf0c319a" (UID: "ebce59c6-74ef-431c-8685-ee7eaf0c319a"). InnerVolumeSpecName "kube-api-access-67qzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.612768 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67qzk\" (UniqueName: \"kubernetes.io/projected/ebce59c6-74ef-431c-8685-ee7eaf0c319a-kube-api-access-67qzk\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.925831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d30d-account-create-cfnvj" event={"ID":"ebce59c6-74ef-431c-8685-ee7eaf0c319a","Type":"ContainerDied","Data":"f153f06182686ba4fcdab31b10990cac7ca989a848c0e15d3005fed432e4a8e1"} Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.925882 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d30d-account-create-cfnvj" Oct 03 15:15:40 crc kubenswrapper[4959]: I1003 15:15:40.925885 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f153f06182686ba4fcdab31b10990cac7ca989a848c0e15d3005fed432e4a8e1" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.124255 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-zxwhz"] Oct 03 15:15:42 crc kubenswrapper[4959]: E1003 15:15:42.125083 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebce59c6-74ef-431c-8685-ee7eaf0c319a" containerName="mariadb-account-create" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.125104 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebce59c6-74ef-431c-8685-ee7eaf0c319a" containerName="mariadb-account-create" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.125410 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebce59c6-74ef-431c-8685-ee7eaf0c319a" containerName="mariadb-account-create" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.126165 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.128540 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.128964 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5lp2l" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.133555 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-zxwhz"] Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.247027 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g75k8\" (UniqueName: \"kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.247263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.247391 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.349550 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g75k8\" (UniqueName: \"kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.349675 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.349722 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.355616 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.359335 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.370481 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g75k8\" (UniqueName: \"kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8\") pod \"heat-db-sync-zxwhz\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.452305 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.934822 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-zxwhz"] Oct 03 15:15:42 crc kubenswrapper[4959]: I1003 15:15:42.939990 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:15:43 crc kubenswrapper[4959]: I1003 15:15:43.955678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-zxwhz" event={"ID":"082b2d7d-c1c4-4916-bcd2-55427433e41b","Type":"ContainerStarted","Data":"6c9124dfff496e7c17d85e7ceb5c7b5418137b6e5e11e0f1e09cb6001edcf39f"} Oct 03 15:15:45 crc kubenswrapper[4959]: I1003 15:15:45.981629 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5656c49bff-fgjq5" podUID="8ce9411f-dcdd-4538-a83e-193382cda8e3" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.121:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8080: connect: connection refused" Oct 03 15:15:50 crc kubenswrapper[4959]: I1003 15:15:50.030046 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-zxwhz" event={"ID":"082b2d7d-c1c4-4916-bcd2-55427433e41b","Type":"ContainerStarted","Data":"84f60ef4f074f7ee031d1ec3c5a017b68b367b8281bee39f60e53e8143cf296e"} Oct 03 15:15:52 crc kubenswrapper[4959]: I1003 15:15:52.052781 4959 generic.go:334] "Generic (PLEG): container finished" podID="082b2d7d-c1c4-4916-bcd2-55427433e41b" containerID="84f60ef4f074f7ee031d1ec3c5a017b68b367b8281bee39f60e53e8143cf296e" exitCode=0 Oct 03 15:15:52 crc kubenswrapper[4959]: I1003 15:15:52.052872 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-zxwhz" event={"ID":"082b2d7d-c1c4-4916-bcd2-55427433e41b","Type":"ContainerDied","Data":"84f60ef4f074f7ee031d1ec3c5a017b68b367b8281bee39f60e53e8143cf296e"} Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.442737 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.574406 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g75k8\" (UniqueName: \"kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8\") pod \"082b2d7d-c1c4-4916-bcd2-55427433e41b\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.575758 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle\") pod \"082b2d7d-c1c4-4916-bcd2-55427433e41b\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.575905 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data\") pod \"082b2d7d-c1c4-4916-bcd2-55427433e41b\" (UID: \"082b2d7d-c1c4-4916-bcd2-55427433e41b\") " Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.581817 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8" (OuterVolumeSpecName: "kube-api-access-g75k8") pod "082b2d7d-c1c4-4916-bcd2-55427433e41b" (UID: "082b2d7d-c1c4-4916-bcd2-55427433e41b"). InnerVolumeSpecName "kube-api-access-g75k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.606408 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "082b2d7d-c1c4-4916-bcd2-55427433e41b" (UID: "082b2d7d-c1c4-4916-bcd2-55427433e41b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.658032 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data" (OuterVolumeSpecName: "config-data") pod "082b2d7d-c1c4-4916-bcd2-55427433e41b" (UID: "082b2d7d-c1c4-4916-bcd2-55427433e41b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.681073 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.681132 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g75k8\" (UniqueName: \"kubernetes.io/projected/082b2d7d-c1c4-4916-bcd2-55427433e41b-kube-api-access-g75k8\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:53 crc kubenswrapper[4959]: I1003 15:15:53.681150 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/082b2d7d-c1c4-4916-bcd2-55427433e41b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:15:54 crc kubenswrapper[4959]: I1003 15:15:54.073045 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-zxwhz" event={"ID":"082b2d7d-c1c4-4916-bcd2-55427433e41b","Type":"ContainerDied","Data":"6c9124dfff496e7c17d85e7ceb5c7b5418137b6e5e11e0f1e09cb6001edcf39f"} Oct 03 15:15:54 crc kubenswrapper[4959]: I1003 15:15:54.073094 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c9124dfff496e7c17d85e7ceb5c7b5418137b6e5e11e0f1e09cb6001edcf39f" Oct 03 15:15:54 crc kubenswrapper[4959]: I1003 15:15:54.073434 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-zxwhz" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.053554 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vdxhq"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.070304 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vdxhq"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.614740 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-67486f58d7-k7xmb"] Oct 03 15:15:55 crc kubenswrapper[4959]: E1003 15:15:55.615588 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082b2d7d-c1c4-4916-bcd2-55427433e41b" containerName="heat-db-sync" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.615610 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="082b2d7d-c1c4-4916-bcd2-55427433e41b" containerName="heat-db-sync" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.615854 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="082b2d7d-c1c4-4916-bcd2-55427433e41b" containerName="heat-db-sync" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.616786 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.634318 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.634394 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.634909 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5lp2l" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.650314 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67486f58d7-k7xmb"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.705867 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5135668-739c-49a6-a7e5-645acd5e8021" path="/var/lib/kubelet/pods/e5135668-739c-49a6-a7e5-645acd5e8021/volumes" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.740671 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data-custom\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.740722 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.740869 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgt2m\" (UniqueName: \"kubernetes.io/projected/7aa03178-4f69-410a-9a12-62bdbf7c556d-kube-api-access-pgt2m\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.740903 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-combined-ca-bundle\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.767172 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7fc99996f4-zzpdp"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.768950 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.770611 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.787513 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7fc99996f4-zzpdp"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.819480 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-799957bf96-8r76b"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.820715 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.825212 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.843556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgt2m\" (UniqueName: \"kubernetes.io/projected/7aa03178-4f69-410a-9a12-62bdbf7c556d-kube-api-access-pgt2m\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.843594 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-combined-ca-bundle\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.843656 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data-custom\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.843693 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.852933 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-799957bf96-8r76b"] Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.854539 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data-custom\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.855083 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-combined-ca-bundle\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.855356 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa03178-4f69-410a-9a12-62bdbf7c556d-config-data\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.880374 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgt2m\" (UniqueName: \"kubernetes.io/projected/7aa03178-4f69-410a-9a12-62bdbf7c556d-kube-api-access-pgt2m\") pod \"heat-engine-67486f58d7-k7xmb\" (UID: \"7aa03178-4f69-410a-9a12-62bdbf7c556d\") " pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.946323 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.946398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-combined-ca-bundle\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.946560 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvhvf\" (UniqueName: \"kubernetes.io/projected/e0417f0e-e43d-4ade-83f6-189961bdf2a9-kube-api-access-mvhvf\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.946797 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.946871 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data-custom\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.947067 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data-custom\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.947369 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-combined-ca-bundle\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.947456 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkwc9\" (UniqueName: \"kubernetes.io/projected/5fcd2bf1-cf03-4438-930d-d87923613b4c-kube-api-access-rkwc9\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:55 crc kubenswrapper[4959]: I1003 15:15:55.971916 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.049340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvhvf\" (UniqueName: \"kubernetes.io/projected/e0417f0e-e43d-4ade-83f6-189961bdf2a9-kube-api-access-mvhvf\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.049636 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050109 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data-custom\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050146 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data-custom\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050406 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-combined-ca-bundle\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050434 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkwc9\" (UniqueName: \"kubernetes.io/projected/5fcd2bf1-cf03-4438-930d-d87923613b4c-kube-api-access-rkwc9\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050477 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.050508 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-combined-ca-bundle\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.054353 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.056416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-config-data-custom\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.078490 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-combined-ca-bundle\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.080043 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.081466 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0417f0e-e43d-4ade-83f6-189961bdf2a9-combined-ca-bundle\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.085089 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fcd2bf1-cf03-4438-930d-d87923613b4c-config-data-custom\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.093187 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkwc9\" (UniqueName: \"kubernetes.io/projected/5fcd2bf1-cf03-4438-930d-d87923613b4c-kube-api-access-rkwc9\") pod \"heat-cfnapi-7fc99996f4-zzpdp\" (UID: \"5fcd2bf1-cf03-4438-930d-d87923613b4c\") " pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.094818 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvhvf\" (UniqueName: \"kubernetes.io/projected/e0417f0e-e43d-4ade-83f6-189961bdf2a9-kube-api-access-mvhvf\") pod \"heat-api-799957bf96-8r76b\" (UID: \"e0417f0e-e43d-4ade-83f6-189961bdf2a9\") " pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.141262 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.258541 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.499645 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-67486f58d7-k7xmb"] Oct 03 15:15:56 crc kubenswrapper[4959]: W1003 15:15:56.529338 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7aa03178_4f69_410a_9a12_62bdbf7c556d.slice/crio-72fef6f3d393c1a7a4d3d2864b34bada8a866bae48fa8f3e1a922fce74af800d WatchSource:0}: Error finding container 72fef6f3d393c1a7a4d3d2864b34bada8a866bae48fa8f3e1a922fce74af800d: Status 404 returned error can't find the container with id 72fef6f3d393c1a7a4d3d2864b34bada8a866bae48fa8f3e1a922fce74af800d Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.693018 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7fc99996f4-zzpdp"] Oct 03 15:15:56 crc kubenswrapper[4959]: I1003 15:15:56.910942 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-799957bf96-8r76b"] Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.127855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" event={"ID":"5fcd2bf1-cf03-4438-930d-d87923613b4c","Type":"ContainerStarted","Data":"6a7f7073811fb0ac55a59d8581c5136508571ef3ed1d7f67b2e9b7e62ef120cc"} Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.130888 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-799957bf96-8r76b" event={"ID":"e0417f0e-e43d-4ade-83f6-189961bdf2a9","Type":"ContainerStarted","Data":"23fade663ef603ef32a0b4e44447ee6d2c0222899a5ff1724ec93bf443176012"} Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.133222 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67486f58d7-k7xmb" event={"ID":"7aa03178-4f69-410a-9a12-62bdbf7c556d","Type":"ContainerStarted","Data":"f3d0ffb0d5bed763108d45b5392578055df6efca68e5d3ae064a4c5004995251"} Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.133261 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-67486f58d7-k7xmb" event={"ID":"7aa03178-4f69-410a-9a12-62bdbf7c556d","Type":"ContainerStarted","Data":"72fef6f3d393c1a7a4d3d2864b34bada8a866bae48fa8f3e1a922fce74af800d"} Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.134525 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:15:57 crc kubenswrapper[4959]: I1003 15:15:57.165226 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-67486f58d7-k7xmb" podStartSLOduration=2.165207943 podStartE2EDuration="2.165207943s" podCreationTimestamp="2025-10-03 15:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:15:57.153741213 +0000 UTC m=+6326.357084630" watchObservedRunningTime="2025-10-03 15:15:57.165207943 +0000 UTC m=+6326.368551360" Oct 03 15:15:58 crc kubenswrapper[4959]: I1003 15:15:58.179414 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.162231 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" event={"ID":"5fcd2bf1-cf03-4438-930d-d87923613b4c","Type":"ContainerStarted","Data":"7c09bd4de1c96a9f095d6c094473d71de40ad4de0529f244d460cd9aa5a3cf93"} Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.162636 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.165569 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-799957bf96-8r76b" event={"ID":"e0417f0e-e43d-4ade-83f6-189961bdf2a9","Type":"ContainerStarted","Data":"86d0efeefacabd1d7f3026dce2561336fd1c7d439dadba03ae5855a899c2282a"} Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.165638 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.190226 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" podStartSLOduration=2.653061283 podStartE2EDuration="5.190203781s" podCreationTimestamp="2025-10-03 15:15:55 +0000 UTC" firstStartedPulling="2025-10-03 15:15:56.726431166 +0000 UTC m=+6325.929774583" lastFinishedPulling="2025-10-03 15:15:59.263573664 +0000 UTC m=+6328.466917081" observedRunningTime="2025-10-03 15:16:00.189121004 +0000 UTC m=+6329.392464421" watchObservedRunningTime="2025-10-03 15:16:00.190203781 +0000 UTC m=+6329.393547198" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.209757 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-799957bf96-8r76b" podStartSLOduration=2.862022987 podStartE2EDuration="5.209743579s" podCreationTimestamp="2025-10-03 15:15:55 +0000 UTC" firstStartedPulling="2025-10-03 15:15:56.918997709 +0000 UTC m=+6326.122341126" lastFinishedPulling="2025-10-03 15:15:59.266718301 +0000 UTC m=+6328.470061718" observedRunningTime="2025-10-03 15:16:00.204001989 +0000 UTC m=+6329.407345406" watchObservedRunningTime="2025-10-03 15:16:00.209743579 +0000 UTC m=+6329.413086996" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.419060 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5656c49bff-fgjq5" Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.480702 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.480920 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon-log" containerID="cri-o://547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a" gracePeriod=30 Oct 03 15:16:00 crc kubenswrapper[4959]: I1003 15:16:00.481064 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" containerID="cri-o://9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0" gracePeriod=30 Oct 03 15:16:04 crc kubenswrapper[4959]: I1003 15:16:04.202249 4959 generic.go:334] "Generic (PLEG): container finished" podID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerID="9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0" exitCode=0 Oct 03 15:16:04 crc kubenswrapper[4959]: I1003 15:16:04.202329 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerDied","Data":"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0"} Oct 03 15:16:05 crc kubenswrapper[4959]: I1003 15:16:05.027313 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-569f-account-create-j5b44"] Oct 03 15:16:05 crc kubenswrapper[4959]: I1003 15:16:05.034563 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-569f-account-create-j5b44"] Oct 03 15:16:05 crc kubenswrapper[4959]: I1003 15:16:05.711151 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46315488-a177-4eb1-ab05-e91501e1b11b" path="/var/lib/kubelet/pods/46315488-a177-4eb1-ab05-e91501e1b11b/volumes" Oct 03 15:16:06 crc kubenswrapper[4959]: I1003 15:16:06.044711 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:16:06 crc kubenswrapper[4959]: I1003 15:16:06.044774 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.532334 4959 scope.go:117] "RemoveContainer" containerID="061a55ca478075c7ca9d94bd6d87d1de94f390f1eade79ab396648f02a33b050" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.599188 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7fc99996f4-zzpdp" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.613526 4959 scope.go:117] "RemoveContainer" containerID="cce3cda0e5531ca24833d5aed6642fc04c4a062b0358d1d4d0af87c6b35789be" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.673596 4959 scope.go:117] "RemoveContainer" containerID="f5cae178a0cb97b7ab5fb961af6ffe5aad0714fb7fca08b2e365cb2dc15ac427" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.740369 4959 scope.go:117] "RemoveContainer" containerID="35a0a5560a92ef70bc58ba719dd4f61449e36a1224065ad4743e605d6b124fca" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.754147 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-799957bf96-8r76b" Oct 03 15:16:07 crc kubenswrapper[4959]: I1003 15:16:07.772965 4959 scope.go:117] "RemoveContainer" containerID="0ee5999a6cd6d503eadb5ed5c19b9fea27fea8409722de7618e2fdd61e9683e6" Oct 03 15:16:08 crc kubenswrapper[4959]: I1003 15:16:08.238726 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 03 15:16:12 crc kubenswrapper[4959]: I1003 15:16:12.044758 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-c6pxg"] Oct 03 15:16:12 crc kubenswrapper[4959]: I1003 15:16:12.061028 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-c6pxg"] Oct 03 15:16:13 crc kubenswrapper[4959]: I1003 15:16:13.706597 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d19694-00e8-4ad4-93b9-e016048c49bf" path="/var/lib/kubelet/pods/27d19694-00e8-4ad4-93b9-e016048c49bf/volumes" Oct 03 15:16:16 crc kubenswrapper[4959]: I1003 15:16:16.014292 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-67486f58d7-k7xmb" Oct 03 15:16:18 crc kubenswrapper[4959]: I1003 15:16:18.238427 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 03 15:16:22 crc kubenswrapper[4959]: E1003 15:16:22.614864 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:39206->38.129.56.245:39745: write tcp 38.129.56.245:39206->38.129.56.245:39745: write: broken pipe Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.675231 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl"] Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.677550 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.682016 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.701260 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl"] Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.723977 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.724491 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdmfb\" (UniqueName: \"kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.724581 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.825648 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.825726 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.825817 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdmfb\" (UniqueName: \"kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.826131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.827241 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:24 crc kubenswrapper[4959]: I1003 15:16:24.844738 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdmfb\" (UniqueName: \"kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:25 crc kubenswrapper[4959]: I1003 15:16:25.004297 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:25 crc kubenswrapper[4959]: I1003 15:16:25.447838 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl"] Oct 03 15:16:26 crc kubenswrapper[4959]: I1003 15:16:26.440358 4959 generic.go:334] "Generic (PLEG): container finished" podID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerID="c7140a8ea4bb7d01adb28f22c9462b84d3451aeed8ced008580f8d0ff719afa5" exitCode=0 Oct 03 15:16:26 crc kubenswrapper[4959]: I1003 15:16:26.440442 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" event={"ID":"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5","Type":"ContainerDied","Data":"c7140a8ea4bb7d01adb28f22c9462b84d3451aeed8ced008580f8d0ff719afa5"} Oct 03 15:16:26 crc kubenswrapper[4959]: I1003 15:16:26.441068 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" event={"ID":"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5","Type":"ContainerStarted","Data":"31a0a3b09cbc523f16d7fde06fba6a379827c4e7b1570799a3c9948cf9a3243c"} Oct 03 15:16:28 crc kubenswrapper[4959]: I1003 15:16:28.239392 4959 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-cb4d9c449-s24wt" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 03 15:16:28 crc kubenswrapper[4959]: I1003 15:16:28.239805 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:16:29 crc kubenswrapper[4959]: I1003 15:16:29.473646 4959 generic.go:334] "Generic (PLEG): container finished" podID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerID="8c49a03c97311e0f72ac46279807f1c39ec7f6c9b289e8b3866f36fcad6e3d63" exitCode=0 Oct 03 15:16:29 crc kubenswrapper[4959]: I1003 15:16:29.473792 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" event={"ID":"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5","Type":"ContainerDied","Data":"8c49a03c97311e0f72ac46279807f1c39ec7f6c9b289e8b3866f36fcad6e3d63"} Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.497422 4959 generic.go:334] "Generic (PLEG): container finished" podID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerID="c8bacfbc55513da5ddd366bacda50cc5b7198e1f04f7c9eb69f312e2c2584689" exitCode=0 Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.497772 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" event={"ID":"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5","Type":"ContainerDied","Data":"c8bacfbc55513da5ddd366bacda50cc5b7198e1f04f7c9eb69f312e2c2584689"} Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.909447 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.958730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts\") pod \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.958780 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwftq\" (UniqueName: \"kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq\") pod \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.958852 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs\") pod \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.958917 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key\") pod \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.958968 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data\") pod \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\" (UID: \"745adf72-91c0-4fcf-a53b-fc50b0f8f346\") " Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.959963 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs" (OuterVolumeSpecName: "logs") pod "745adf72-91c0-4fcf-a53b-fc50b0f8f346" (UID: "745adf72-91c0-4fcf-a53b-fc50b0f8f346"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.967335 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "745adf72-91c0-4fcf-a53b-fc50b0f8f346" (UID: "745adf72-91c0-4fcf-a53b-fc50b0f8f346"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.967966 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq" (OuterVolumeSpecName: "kube-api-access-zwftq") pod "745adf72-91c0-4fcf-a53b-fc50b0f8f346" (UID: "745adf72-91c0-4fcf-a53b-fc50b0f8f346"). InnerVolumeSpecName "kube-api-access-zwftq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.985392 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts" (OuterVolumeSpecName: "scripts") pod "745adf72-91c0-4fcf-a53b-fc50b0f8f346" (UID: "745adf72-91c0-4fcf-a53b-fc50b0f8f346"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:16:30 crc kubenswrapper[4959]: I1003 15:16:30.987732 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data" (OuterVolumeSpecName: "config-data") pod "745adf72-91c0-4fcf-a53b-fc50b0f8f346" (UID: "745adf72-91c0-4fcf-a53b-fc50b0f8f346"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.061748 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.061784 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwftq\" (UniqueName: \"kubernetes.io/projected/745adf72-91c0-4fcf-a53b-fc50b0f8f346-kube-api-access-zwftq\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.061798 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/745adf72-91c0-4fcf-a53b-fc50b0f8f346-logs\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.061809 4959 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/745adf72-91c0-4fcf-a53b-fc50b0f8f346-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.061821 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/745adf72-91c0-4fcf-a53b-fc50b0f8f346-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.512547 4959 generic.go:334] "Generic (PLEG): container finished" podID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerID="547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a" exitCode=137 Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.513348 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cb4d9c449-s24wt" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.513339 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerDied","Data":"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a"} Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.514646 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cb4d9c449-s24wt" event={"ID":"745adf72-91c0-4fcf-a53b-fc50b0f8f346","Type":"ContainerDied","Data":"22d488197243bcacfaf50808d2b4ad524bbd8d1ae1edf133c2eb9056c569511e"} Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.514672 4959 scope.go:117] "RemoveContainer" containerID="9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.562347 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.575674 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-cb4d9c449-s24wt"] Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.711251 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" path="/var/lib/kubelet/pods/745adf72-91c0-4fcf-a53b-fc50b0f8f346/volumes" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.795959 4959 scope.go:117] "RemoveContainer" containerID="547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.884568 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.913019 4959 scope.go:117] "RemoveContainer" containerID="9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0" Oct 03 15:16:31 crc kubenswrapper[4959]: E1003 15:16:31.913468 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0\": container with ID starting with 9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0 not found: ID does not exist" containerID="9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.913518 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0"} err="failed to get container status \"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0\": rpc error: code = NotFound desc = could not find container \"9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0\": container with ID starting with 9d86ea4f2be98724938219748fe26b9023e3b4c2f1c914293c7d4fff116352d0 not found: ID does not exist" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.913544 4959 scope.go:117] "RemoveContainer" containerID="547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a" Oct 03 15:16:31 crc kubenswrapper[4959]: E1003 15:16:31.917017 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a\": container with ID starting with 547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a not found: ID does not exist" containerID="547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.917068 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a"} err="failed to get container status \"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a\": rpc error: code = NotFound desc = could not find container \"547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a\": container with ID starting with 547ed609e0fc1331cccd9f0a485e0ca482b4d5646912aae56e3ea3ecdfbe315a not found: ID does not exist" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.986145 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdmfb\" (UniqueName: \"kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb\") pod \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.986253 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util\") pod \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.986422 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle\") pod \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\" (UID: \"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5\") " Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.988184 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle" (OuterVolumeSpecName: "bundle") pod "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" (UID: "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.990756 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb" (OuterVolumeSpecName: "kube-api-access-xdmfb") pod "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" (UID: "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5"). InnerVolumeSpecName "kube-api-access-xdmfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:16:31 crc kubenswrapper[4959]: I1003 15:16:31.996939 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util" (OuterVolumeSpecName: "util") pod "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" (UID: "b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.089092 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdmfb\" (UniqueName: \"kubernetes.io/projected/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-kube-api-access-xdmfb\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.089127 4959 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-util\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.089139 4959 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.525171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" event={"ID":"b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5","Type":"ContainerDied","Data":"31a0a3b09cbc523f16d7fde06fba6a379827c4e7b1570799a3c9948cf9a3243c"} Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.525231 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31a0a3b09cbc523f16d7fde06fba6a379827c4e7b1570799a3c9948cf9a3243c" Oct 03 15:16:32 crc kubenswrapper[4959]: I1003 15:16:32.525237 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.044292 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.044757 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.044800 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.045586 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.045638 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" gracePeriod=600 Oct 03 15:16:36 crc kubenswrapper[4959]: E1003 15:16:36.181370 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.566294 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" exitCode=0 Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.566336 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176"} Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.566374 4959 scope.go:117] "RemoveContainer" containerID="daedc432e2ec15ee12f92a00701343c0b0c7290379bf0c8a84a95ae41332ff2a" Oct 03 15:16:36 crc kubenswrapper[4959]: I1003 15:16:36.567168 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:16:36 crc kubenswrapper[4959]: E1003 15:16:36.567523 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.715564 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh"] Oct 03 15:16:41 crc kubenswrapper[4959]: E1003 15:16:41.717937 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.718059 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" Oct 03 15:16:41 crc kubenswrapper[4959]: E1003 15:16:41.718165 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="extract" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.718299 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="extract" Oct 03 15:16:41 crc kubenswrapper[4959]: E1003 15:16:41.718378 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="pull" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.718450 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="pull" Oct 03 15:16:41 crc kubenswrapper[4959]: E1003 15:16:41.718544 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="util" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.718614 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="util" Oct 03 15:16:41 crc kubenswrapper[4959]: E1003 15:16:41.718693 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon-log" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.718760 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon-log" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.719116 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.719236 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="745adf72-91c0-4fcf-a53b-fc50b0f8f346" containerName="horizon-log" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.719333 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5" containerName="extract" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.721876 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.728094 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.728310 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-drkjd" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.728795 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.739632 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.797590 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.798951 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.805025 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.805275 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-q67jh" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.825833 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.828459 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.842610 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.866660 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.912070 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v85pc\" (UniqueName: \"kubernetes.io/projected/35ba318c-c0b5-41fc-9ca6-a038d6a43622-kube-api-access-v85pc\") pod \"obo-prometheus-operator-7c8cf85677-9cvxh\" (UID: \"35ba318c-c0b5-41fc-9ca6-a038d6a43622\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.912139 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.912325 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.955172 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-s5zlt"] Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.956381 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.957881 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 03 15:16:41 crc kubenswrapper[4959]: I1003 15:16:41.958122 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-jrdtd" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.017332 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.017427 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.017455 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.017515 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v85pc\" (UniqueName: \"kubernetes.io/projected/35ba318c-c0b5-41fc-9ca6-a038d6a43622-kube-api-access-v85pc\") pod \"obo-prometheus-operator-7c8cf85677-9cvxh\" (UID: \"35ba318c-c0b5-41fc-9ca6-a038d6a43622\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.017563 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.024306 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.025866 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/77daaec5-27eb-430e-8f11-fa6e89b06589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn\" (UID: \"77daaec5-27eb-430e-8f11-fa6e89b06589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.038393 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v85pc\" (UniqueName: \"kubernetes.io/projected/35ba318c-c0b5-41fc-9ca6-a038d6a43622-kube-api-access-v85pc\") pod \"obo-prometheus-operator-7c8cf85677-9cvxh\" (UID: \"35ba318c-c0b5-41fc-9ca6-a038d6a43622\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.038961 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-s5zlt"] Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.053507 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.119168 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/19da9e3c-bf60-43de-965e-35faf0b981d4-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.119320 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.119354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.119415 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnc6m\" (UniqueName: \"kubernetes.io/projected/19da9e3c-bf60-43de-965e-35faf0b981d4-kube-api-access-tnc6m\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.123728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.126830 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d4bd245c-1646-4475-a8e9-b07a581ea59e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq\" (UID: \"d4bd245c-1646-4475-a8e9-b07a581ea59e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.132671 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.142531 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x4s2b"] Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.145708 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.153094 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-2p7nd" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.153756 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x4s2b"] Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.181727 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.222235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnc6m\" (UniqueName: \"kubernetes.io/projected/19da9e3c-bf60-43de-965e-35faf0b981d4-kube-api-access-tnc6m\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.222337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/19da9e3c-bf60-43de-965e-35faf0b981d4-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.225802 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/19da9e3c-bf60-43de-965e-35faf0b981d4-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.244322 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnc6m\" (UniqueName: \"kubernetes.io/projected/19da9e3c-bf60-43de-965e-35faf0b981d4-kube-api-access-tnc6m\") pod \"observability-operator-cc5f78dfc-s5zlt\" (UID: \"19da9e3c-bf60-43de-965e-35faf0b981d4\") " pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.311891 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.323903 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtxxj\" (UniqueName: \"kubernetes.io/projected/448da875-3b09-4c56-ae51-cb99e2483b30-kube-api-access-qtxxj\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.323959 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/448da875-3b09-4c56-ae51-cb99e2483b30-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.425468 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtxxj\" (UniqueName: \"kubernetes.io/projected/448da875-3b09-4c56-ae51-cb99e2483b30-kube-api-access-qtxxj\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.425521 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/448da875-3b09-4c56-ae51-cb99e2483b30-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.426976 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/448da875-3b09-4c56-ae51-cb99e2483b30-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.446453 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtxxj\" (UniqueName: \"kubernetes.io/projected/448da875-3b09-4c56-ae51-cb99e2483b30-kube-api-access-qtxxj\") pod \"perses-operator-54bc95c9fb-x4s2b\" (UID: \"448da875-3b09-4c56-ae51-cb99e2483b30\") " pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.594865 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.712084 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh"] Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.790602 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq"] Oct 03 15:16:42 crc kubenswrapper[4959]: W1003 15:16:42.814691 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4bd245c_1646_4475_a8e9_b07a581ea59e.slice/crio-47dd9df92fa79a358ecf2cd27aacd478a540316f50b71252fec5a6766d112095 WatchSource:0}: Error finding container 47dd9df92fa79a358ecf2cd27aacd478a540316f50b71252fec5a6766d112095: Status 404 returned error can't find the container with id 47dd9df92fa79a358ecf2cd27aacd478a540316f50b71252fec5a6766d112095 Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.862362 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn"] Oct 03 15:16:42 crc kubenswrapper[4959]: I1003 15:16:42.920860 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-s5zlt"] Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.247616 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x4s2b"] Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.678607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" event={"ID":"19da9e3c-bf60-43de-965e-35faf0b981d4","Type":"ContainerStarted","Data":"e9a22098087f01e641fa92f605b311434a066673116bb62e130693582574f0c4"} Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.680911 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" event={"ID":"77daaec5-27eb-430e-8f11-fa6e89b06589","Type":"ContainerStarted","Data":"4f034bec693af6f2b798170707c037bf9b265876c6f8d2b7c0f83a5fea1fc037"} Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.682699 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" event={"ID":"448da875-3b09-4c56-ae51-cb99e2483b30","Type":"ContainerStarted","Data":"c77a68c4c459ff87ad1b760b9e8546d15c1336a41ac2514f7a4345f5d04444d9"} Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.684211 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" event={"ID":"d4bd245c-1646-4475-a8e9-b07a581ea59e","Type":"ContainerStarted","Data":"47dd9df92fa79a358ecf2cd27aacd478a540316f50b71252fec5a6766d112095"} Oct 03 15:16:43 crc kubenswrapper[4959]: I1003 15:16:43.696907 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" event={"ID":"35ba318c-c0b5-41fc-9ca6-a038d6a43622","Type":"ContainerStarted","Data":"6cd05abf272b1c833bc55ae1e9f8b418101c4dc1016154a0de1d191a988a537f"} Oct 03 15:16:46 crc kubenswrapper[4959]: I1003 15:16:46.718932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" event={"ID":"d4bd245c-1646-4475-a8e9-b07a581ea59e","Type":"ContainerStarted","Data":"02c1f2514cca4229a2b70e4e16be2ee28abb153e2600ec74c130e9960c3913b9"} Oct 03 15:16:46 crc kubenswrapper[4959]: I1003 15:16:46.720160 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" event={"ID":"77daaec5-27eb-430e-8f11-fa6e89b06589","Type":"ContainerStarted","Data":"c6b634cac587464ba86de7f4b3272f7465b687560627333e98b77ee5cb5fb106"} Oct 03 15:16:46 crc kubenswrapper[4959]: I1003 15:16:46.754479 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq" podStartSLOduration=3.011359583 podStartE2EDuration="5.754460981s" podCreationTimestamp="2025-10-03 15:16:41 +0000 UTC" firstStartedPulling="2025-10-03 15:16:42.910059601 +0000 UTC m=+6372.113403018" lastFinishedPulling="2025-10-03 15:16:45.653160999 +0000 UTC m=+6374.856504416" observedRunningTime="2025-10-03 15:16:46.746400743 +0000 UTC m=+6375.949744160" watchObservedRunningTime="2025-10-03 15:16:46.754460981 +0000 UTC m=+6375.957804388" Oct 03 15:16:49 crc kubenswrapper[4959]: I1003 15:16:49.780914 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" event={"ID":"448da875-3b09-4c56-ae51-cb99e2483b30","Type":"ContainerStarted","Data":"b5ef7ebc0d7a13e7f470c638f455ea28be5ceed7ba32f99afa21dcf65ad964f3"} Oct 03 15:16:49 crc kubenswrapper[4959]: I1003 15:16:49.784868 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:16:49 crc kubenswrapper[4959]: I1003 15:16:49.808762 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn" podStartSLOduration=6.066972129 podStartE2EDuration="8.808747245s" podCreationTimestamp="2025-10-03 15:16:41 +0000 UTC" firstStartedPulling="2025-10-03 15:16:42.9096177 +0000 UTC m=+6372.112961117" lastFinishedPulling="2025-10-03 15:16:45.651392816 +0000 UTC m=+6374.854736233" observedRunningTime="2025-10-03 15:16:46.771720183 +0000 UTC m=+6375.975063600" watchObservedRunningTime="2025-10-03 15:16:49.808747245 +0000 UTC m=+6379.012090662" Oct 03 15:16:49 crc kubenswrapper[4959]: I1003 15:16:49.818244 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" podStartSLOduration=1.845910082 podStartE2EDuration="7.818223077s" podCreationTimestamp="2025-10-03 15:16:42 +0000 UTC" firstStartedPulling="2025-10-03 15:16:43.253510455 +0000 UTC m=+6372.456853872" lastFinishedPulling="2025-10-03 15:16:49.22582345 +0000 UTC m=+6378.429166867" observedRunningTime="2025-10-03 15:16:49.809241077 +0000 UTC m=+6379.012584494" watchObservedRunningTime="2025-10-03 15:16:49.818223077 +0000 UTC m=+6379.021566494" Oct 03 15:16:51 crc kubenswrapper[4959]: I1003 15:16:51.697892 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:16:51 crc kubenswrapper[4959]: E1003 15:16:51.698374 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:16:52 crc kubenswrapper[4959]: I1003 15:16:52.841885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" event={"ID":"19da9e3c-bf60-43de-965e-35faf0b981d4","Type":"ContainerStarted","Data":"db3db2bbedb0e16ce1b90b31f9eb23612a0e25a2ec97ab09bbc9acf2b350d2ab"} Oct 03 15:16:52 crc kubenswrapper[4959]: I1003 15:16:52.842518 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:52 crc kubenswrapper[4959]: I1003 15:16:52.844473 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" Oct 03 15:16:52 crc kubenswrapper[4959]: I1003 15:16:52.867333 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-s5zlt" podStartSLOduration=2.479176428 podStartE2EDuration="11.867305764s" podCreationTimestamp="2025-10-03 15:16:41 +0000 UTC" firstStartedPulling="2025-10-03 15:16:42.910643505 +0000 UTC m=+6372.113986922" lastFinishedPulling="2025-10-03 15:16:52.298772821 +0000 UTC m=+6381.502116258" observedRunningTime="2025-10-03 15:16:52.864470925 +0000 UTC m=+6382.067814372" watchObservedRunningTime="2025-10-03 15:16:52.867305764 +0000 UTC m=+6382.070649201" Oct 03 15:16:59 crc kubenswrapper[4959]: I1003 15:16:59.910754 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" event={"ID":"35ba318c-c0b5-41fc-9ca6-a038d6a43622","Type":"ContainerStarted","Data":"858773b58af4735a0be2826de360d8b935cd14988c2ecd3a5d5344eae88e283c"} Oct 03 15:16:59 crc kubenswrapper[4959]: I1003 15:16:59.933231 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-9cvxh" podStartSLOduration=2.981310826 podStartE2EDuration="18.93321036s" podCreationTimestamp="2025-10-03 15:16:41 +0000 UTC" firstStartedPulling="2025-10-03 15:16:42.727986715 +0000 UTC m=+6371.931330132" lastFinishedPulling="2025-10-03 15:16:58.679886249 +0000 UTC m=+6387.883229666" observedRunningTime="2025-10-03 15:16:59.925947123 +0000 UTC m=+6389.129290560" watchObservedRunningTime="2025-10-03 15:16:59.93321036 +0000 UTC m=+6389.136553787" Oct 03 15:17:02 crc kubenswrapper[4959]: I1003 15:17:02.598436 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-x4s2b" Oct 03 15:17:02 crc kubenswrapper[4959]: I1003 15:17:02.688076 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:17:02 crc kubenswrapper[4959]: E1003 15:17:02.688381 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.486788 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.487242 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" containerName="openstackclient" containerID="cri-o://c7ea5b66e58c444ff1381a4fcf5bfc87affa757002b29fff1d266e600f74daaa" gracePeriod=2 Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.495304 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.556366 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 15:17:05 crc kubenswrapper[4959]: E1003 15:17:05.556936 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" containerName="openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.557006 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" containerName="openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.557299 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" containerName="openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.558014 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.582568 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.585584 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" podUID="236209a8-8d65-4c3f-9803-824fca99dd36" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.664624 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.664674 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgcpg\" (UniqueName: \"kubernetes.io/projected/236209a8-8d65-4c3f-9803-824fca99dd36-kube-api-access-cgcpg\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.664706 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.708781 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.709973 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.719898 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.727523 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gzbcs" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.769493 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.769535 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgcpg\" (UniqueName: \"kubernetes.io/projected/236209a8-8d65-4c3f-9803-824fca99dd36-kube-api-access-cgcpg\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.769573 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.770733 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.779828 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/236209a8-8d65-4c3f-9803-824fca99dd36-openstack-config-secret\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.805063 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgcpg\" (UniqueName: \"kubernetes.io/projected/236209a8-8d65-4c3f-9803-824fca99dd36-kube-api-access-cgcpg\") pod \"openstackclient\" (UID: \"236209a8-8d65-4c3f-9803-824fca99dd36\") " pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.871453 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvqs6\" (UniqueName: \"kubernetes.io/projected/8ffac68e-7637-40b7-a7fd-951926faac16-kube-api-access-dvqs6\") pod \"kube-state-metrics-0\" (UID: \"8ffac68e-7637-40b7-a7fd-951926faac16\") " pod="openstack/kube-state-metrics-0" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.911839 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:17:05 crc kubenswrapper[4959]: I1003 15:17:05.975208 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvqs6\" (UniqueName: \"kubernetes.io/projected/8ffac68e-7637-40b7-a7fd-951926faac16-kube-api-access-dvqs6\") pod \"kube-state-metrics-0\" (UID: \"8ffac68e-7637-40b7-a7fd-951926faac16\") " pod="openstack/kube-state-metrics-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.019524 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvqs6\" (UniqueName: \"kubernetes.io/projected/8ffac68e-7637-40b7-a7fd-951926faac16-kube-api-access-dvqs6\") pod \"kube-state-metrics-0\" (UID: \"8ffac68e-7637-40b7-a7fd-951926faac16\") " pod="openstack/kube-state-metrics-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.039746 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.685902 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.692467 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.701286 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.702721 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-mkb52" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.702839 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.702893 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.703045 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.810735 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.810790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.810840 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.810910 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvlf\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-kube-api-access-cqvlf\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.810997 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.811061 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.914708 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.914814 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.915009 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.915062 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.915742 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.915800 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.915909 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvlf\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-kube-api-access-cqvlf\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.938229 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.941779 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.943003 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.944327 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17939c3e-ae90-4f2f-983b-7f4011b95ef2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:06 crc kubenswrapper[4959]: I1003 15:17:06.948418 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvlf\" (UniqueName: \"kubernetes.io/projected/17939c3e-ae90-4f2f-983b-7f4011b95ef2-kube-api-access-cqvlf\") pod \"alertmanager-metric-storage-0\" (UID: \"17939c3e-ae90-4f2f-983b-7f4011b95ef2\") " pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.036742 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.051973 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.218878 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.321618 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.323944 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336381 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336424 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-sm29z" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336539 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336612 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336807 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.336911 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.343219 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.439827 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440127 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440171 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zftb2\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-kube-api-access-zftb2\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440239 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440356 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440384 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.440409 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.545851 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zftb2\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-kube-api-access-zftb2\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.545936 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546035 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546071 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546128 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546214 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.546249 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.547618 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.551058 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.553941 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.554413 4959 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.554451 4959 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/728fcf13b5cb583b479562cc6b9861870d396dd03da3cb65ef6fd65df4463ac5/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.554551 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.556357 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.561877 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.574344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zftb2\" (UniqueName: \"kubernetes.io/projected/fb8fcdfb-38cd-4922-91e9-b0a3be243d78-kube-api-access-zftb2\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.632452 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5560c605-e469-4b4e-a3ec-1b36cc4a840b\") pod \"prometheus-metric-storage-0\" (UID: \"fb8fcdfb-38cd-4922-91e9-b0a3be243d78\") " pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.731931 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.732807 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 03 15:17:07 crc kubenswrapper[4959]: W1003 15:17:07.769217 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17939c3e_ae90_4f2f_983b_7f4011b95ef2.slice/crio-3b0dcf74582c5b39092b040777b437e85593d9b558d0063177f27d804693ca4f WatchSource:0}: Error finding container 3b0dcf74582c5b39092b040777b437e85593d9b558d0063177f27d804693ca4f: Status 404 returned error can't find the container with id 3b0dcf74582c5b39092b040777b437e85593d9b558d0063177f27d804693ca4f Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.946810 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:17:07 crc kubenswrapper[4959]: I1003 15:17:07.977601 4959 scope.go:117] "RemoveContainer" containerID="c7ea5b66e58c444ff1381a4fcf5bfc87affa757002b29fff1d266e600f74daaa" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.031753 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-m2q8h"] Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.033338 4959 scope.go:117] "RemoveContainer" containerID="e096bcae9c5073352458aef380dde247924485f3c52922c9d2d4d2f9255bc6d6" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.042036 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-69rpg"] Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.059592 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-69rpg"] Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.060432 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret\") pod \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.060656 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjwp9\" (UniqueName: \"kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9\") pod \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.060795 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config\") pod \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\" (UID: \"24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42\") " Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.070746 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9" (OuterVolumeSpecName: "kube-api-access-kjwp9") pod "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" (UID: "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42"). InnerVolumeSpecName "kube-api-access-kjwp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.086252 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-m2q8h"] Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.119429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"17939c3e-ae90-4f2f-983b-7f4011b95ef2","Type":"ContainerStarted","Data":"3b0dcf74582c5b39092b040777b437e85593d9b558d0063177f27d804693ca4f"} Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.121479 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8ffac68e-7637-40b7-a7fd-951926faac16","Type":"ContainerStarted","Data":"56efae65433ec766820c3f5c892d35f5aa0bd4255d09bfe4d21d691cc3aad633"} Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.121505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8ffac68e-7637-40b7-a7fd-951926faac16","Type":"ContainerStarted","Data":"27a0971d699bea3de16e037b9918c36663ae2f0ba530c91497d4fa2e43a8d88d"} Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.122802 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.140094 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" (UID: "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.154191 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.559229535 podStartE2EDuration="3.154170054s" podCreationTimestamp="2025-10-03 15:17:05 +0000 UTC" firstStartedPulling="2025-10-03 15:17:07.121624365 +0000 UTC m=+6396.324967782" lastFinishedPulling="2025-10-03 15:17:07.716564884 +0000 UTC m=+6396.919908301" observedRunningTime="2025-10-03 15:17:08.141175555 +0000 UTC m=+6397.344518972" watchObservedRunningTime="2025-10-03 15:17:08.154170054 +0000 UTC m=+6397.357513471" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.156367 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.164436 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.164465 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjwp9\" (UniqueName: \"kubernetes.io/projected/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-kube-api-access-kjwp9\") on node \"crc\" DevicePath \"\"" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.165957 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"236209a8-8d65-4c3f-9803-824fca99dd36","Type":"ContainerStarted","Data":"70f7d8ebc3257c211fa56250f927f248a6e3500161d694a26e5fe126fb4f0a4f"} Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.166008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"236209a8-8d65-4c3f-9803-824fca99dd36","Type":"ContainerStarted","Data":"94f904260143ee0435fa432bc4413a83e0337afca4d34d0afc0b1a7b1e0e00db"} Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.191491 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" (UID: "24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.205234 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.205178722 podStartE2EDuration="3.205178722s" podCreationTimestamp="2025-10-03 15:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:17:08.191639451 +0000 UTC m=+6397.394982868" watchObservedRunningTime="2025-10-03 15:17:08.205178722 +0000 UTC m=+6397.408522139" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.266053 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.416158 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 03 15:17:08 crc kubenswrapper[4959]: I1003 15:17:08.473118 4959 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" podUID="236209a8-8d65-4c3f-9803-824fca99dd36" Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.026508 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-qdks2"] Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.037394 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-qdks2"] Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.181102 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerStarted","Data":"300a94fa09f562674b712405ee7e8db45064d102b008696e4fd57f8c8f41eb38"} Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.704045 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42" path="/var/lib/kubelet/pods/24fdeb57-9ef1-4c1f-b770-0ae7bbe26e42/volumes" Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.704824 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d4256b-825f-4ee1-b1de-3d8db5cf6ed5" path="/var/lib/kubelet/pods/45d4256b-825f-4ee1-b1de-3d8db5cf6ed5/volumes" Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.705367 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5015457d-5742-4f2e-82b1-e9da287f6332" path="/var/lib/kubelet/pods/5015457d-5742-4f2e-82b1-e9da287f6332/volumes" Oct 03 15:17:09 crc kubenswrapper[4959]: I1003 15:17:09.706497 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a79b8d66-b0a9-4a18-ae85-574106748f8a" path="/var/lib/kubelet/pods/a79b8d66-b0a9-4a18-ae85-574106748f8a/volumes" Oct 03 15:17:16 crc kubenswrapper[4959]: I1003 15:17:16.047297 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 15:17:16 crc kubenswrapper[4959]: I1003 15:17:16.252132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"17939c3e-ae90-4f2f-983b-7f4011b95ef2","Type":"ContainerStarted","Data":"f550d36b5e50ceb8861d510a98fea28e8c9fc7400048ddcdcf0d8f3d6e599d36"} Oct 03 15:17:16 crc kubenswrapper[4959]: I1003 15:17:16.253780 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerStarted","Data":"3fd025c8ed0d472671bfc5f55b1788c841918c6b6a92671d420706aa2a5630c8"} Oct 03 15:17:16 crc kubenswrapper[4959]: I1003 15:17:16.685520 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:17:16 crc kubenswrapper[4959]: E1003 15:17:16.686080 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:17:18 crc kubenswrapper[4959]: I1003 15:17:18.050783 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-3366-account-create-h2hwt"] Oct 03 15:17:18 crc kubenswrapper[4959]: I1003 15:17:18.063858 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db25-account-create-tccj6"] Oct 03 15:17:18 crc kubenswrapper[4959]: I1003 15:17:18.075875 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-3366-account-create-h2hwt"] Oct 03 15:17:18 crc kubenswrapper[4959]: I1003 15:17:18.109041 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db25-account-create-tccj6"] Oct 03 15:17:19 crc kubenswrapper[4959]: I1003 15:17:19.059536 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b5b8-account-create-4qrx2"] Oct 03 15:17:19 crc kubenswrapper[4959]: I1003 15:17:19.080288 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b5b8-account-create-4qrx2"] Oct 03 15:17:19 crc kubenswrapper[4959]: I1003 15:17:19.699829 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a92e66-c252-4c32-9f83-05d555300a64" path="/var/lib/kubelet/pods/35a92e66-c252-4c32-9f83-05d555300a64/volumes" Oct 03 15:17:19 crc kubenswrapper[4959]: I1003 15:17:19.701747 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac5a899-a2cd-4d75-85a9-fdab2926e695" path="/var/lib/kubelet/pods/7ac5a899-a2cd-4d75-85a9-fdab2926e695/volumes" Oct 03 15:17:19 crc kubenswrapper[4959]: I1003 15:17:19.702654 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3dbff33-0ddf-4691-ae33-4190fa73a648" path="/var/lib/kubelet/pods/f3dbff33-0ddf-4691-ae33-4190fa73a648/volumes" Oct 03 15:17:24 crc kubenswrapper[4959]: I1003 15:17:24.351669 4959 generic.go:334] "Generic (PLEG): container finished" podID="fb8fcdfb-38cd-4922-91e9-b0a3be243d78" containerID="3fd025c8ed0d472671bfc5f55b1788c841918c6b6a92671d420706aa2a5630c8" exitCode=0 Oct 03 15:17:24 crc kubenswrapper[4959]: I1003 15:17:24.351723 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerDied","Data":"3fd025c8ed0d472671bfc5f55b1788c841918c6b6a92671d420706aa2a5630c8"} Oct 03 15:17:26 crc kubenswrapper[4959]: I1003 15:17:26.380149 4959 generic.go:334] "Generic (PLEG): container finished" podID="17939c3e-ae90-4f2f-983b-7f4011b95ef2" containerID="f550d36b5e50ceb8861d510a98fea28e8c9fc7400048ddcdcf0d8f3d6e599d36" exitCode=0 Oct 03 15:17:26 crc kubenswrapper[4959]: I1003 15:17:26.380221 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"17939c3e-ae90-4f2f-983b-7f4011b95ef2","Type":"ContainerDied","Data":"f550d36b5e50ceb8861d510a98fea28e8c9fc7400048ddcdcf0d8f3d6e599d36"} Oct 03 15:17:28 crc kubenswrapper[4959]: I1003 15:17:28.042535 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tbfhj"] Oct 03 15:17:28 crc kubenswrapper[4959]: I1003 15:17:28.054704 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tbfhj"] Oct 03 15:17:29 crc kubenswrapper[4959]: I1003 15:17:29.696697 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a035a96b-0609-470c-9bf1-066f2b76387c" path="/var/lib/kubelet/pods/a035a96b-0609-470c-9bf1-066f2b76387c/volumes" Oct 03 15:17:30 crc kubenswrapper[4959]: I1003 15:17:30.685839 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:17:30 crc kubenswrapper[4959]: E1003 15:17:30.686406 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:17:31 crc kubenswrapper[4959]: I1003 15:17:31.438760 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerStarted","Data":"95b27c4d04ac2d34bbf83caee2aafd0ebd26e4c154ccd877ddfe59b723998b6a"} Oct 03 15:17:32 crc kubenswrapper[4959]: I1003 15:17:32.452760 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"17939c3e-ae90-4f2f-983b-7f4011b95ef2","Type":"ContainerStarted","Data":"df2ee887a2329415da0026662807bfe4b27db18b3a55220b60f90dc34cfc6918"} Oct 03 15:17:34 crc kubenswrapper[4959]: I1003 15:17:34.504923 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerStarted","Data":"e6aae65c182c10091e2b5f8c97b09e60c97e326b819cdd84e7815b76f6277fa4"} Oct 03 15:17:36 crc kubenswrapper[4959]: I1003 15:17:36.527460 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"17939c3e-ae90-4f2f-983b-7f4011b95ef2","Type":"ContainerStarted","Data":"e3fbb0dcb211fc3418c3756acd0527461889010907731ba151a325fa1a2a86e7"} Oct 03 15:17:36 crc kubenswrapper[4959]: I1003 15:17:36.529955 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:36 crc kubenswrapper[4959]: I1003 15:17:36.534254 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 03 15:17:36 crc kubenswrapper[4959]: I1003 15:17:36.553264 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.654542338 podStartE2EDuration="30.553249868s" podCreationTimestamp="2025-10-03 15:17:06 +0000 UTC" firstStartedPulling="2025-10-03 15:17:07.78990109 +0000 UTC m=+6396.993244507" lastFinishedPulling="2025-10-03 15:17:31.68860862 +0000 UTC m=+6420.891952037" observedRunningTime="2025-10-03 15:17:36.551042104 +0000 UTC m=+6425.754385531" watchObservedRunningTime="2025-10-03 15:17:36.553249868 +0000 UTC m=+6425.756593285" Oct 03 15:17:38 crc kubenswrapper[4959]: I1003 15:17:38.552568 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fb8fcdfb-38cd-4922-91e9-b0a3be243d78","Type":"ContainerStarted","Data":"904628159b35fe106686ca1cf9595ce93a660b6befad4bf8c044eabfdc598f94"} Oct 03 15:17:38 crc kubenswrapper[4959]: I1003 15:17:38.593584 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.258237194 podStartE2EDuration="32.593562619s" podCreationTimestamp="2025-10-03 15:17:06 +0000 UTC" firstStartedPulling="2025-10-03 15:17:08.422749947 +0000 UTC m=+6397.626093364" lastFinishedPulling="2025-10-03 15:17:37.758075382 +0000 UTC m=+6426.961418789" observedRunningTime="2025-10-03 15:17:38.576282466 +0000 UTC m=+6427.779625963" watchObservedRunningTime="2025-10-03 15:17:38.593562619 +0000 UTC m=+6427.796906036" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.043420 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.047955 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.050455 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.050455 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.055866 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.140933 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141003 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141090 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k7zf\" (UniqueName: \"kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141138 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141167 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141372 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.141402 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243701 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243757 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243787 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k7zf\" (UniqueName: \"kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243833 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243864 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243919 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.243973 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.244374 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.244820 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.250607 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.251774 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.276561 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.277004 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k7zf\" (UniqueName: \"kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.277247 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data\") pod \"ceilometer-0\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.366387 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.738579 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:42 crc kubenswrapper[4959]: I1003 15:17:42.916002 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:17:42 crc kubenswrapper[4959]: W1003 15:17:42.917994 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1769ee7_ccfd_44d6_8bec_8e3a4697c358.slice/crio-f4afade29a440ccb599219b226ff29ff98337615fb9bb9aff2415b57e233e3c1 WatchSource:0}: Error finding container f4afade29a440ccb599219b226ff29ff98337615fb9bb9aff2415b57e233e3c1: Status 404 returned error can't find the container with id f4afade29a440ccb599219b226ff29ff98337615fb9bb9aff2415b57e233e3c1 Oct 03 15:17:43 crc kubenswrapper[4959]: I1003 15:17:43.629805 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerStarted","Data":"f4afade29a440ccb599219b226ff29ff98337615fb9bb9aff2415b57e233e3c1"} Oct 03 15:17:44 crc kubenswrapper[4959]: I1003 15:17:44.641572 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerStarted","Data":"5a7e1166a0dcd3e69a8ad07b9b7a695b1625ff9044ea1bb8891a48de902d44b7"} Oct 03 15:17:45 crc kubenswrapper[4959]: I1003 15:17:45.654468 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerStarted","Data":"6067fa1dd65d9a134141290137b46517df72da5cf669202fd9908ecda943dbff"} Oct 03 15:17:45 crc kubenswrapper[4959]: I1003 15:17:45.697712 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:17:45 crc kubenswrapper[4959]: E1003 15:17:45.698294 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:17:46 crc kubenswrapper[4959]: I1003 15:17:46.041570 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gmgst"] Oct 03 15:17:46 crc kubenswrapper[4959]: I1003 15:17:46.048975 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gmgst"] Oct 03 15:17:46 crc kubenswrapper[4959]: I1003 15:17:46.666600 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerStarted","Data":"d27d367a5d3094456370467e4e26062d8c5819a48f9efaa0bdbf1bca5289c251"} Oct 03 15:17:47 crc kubenswrapper[4959]: I1003 15:17:47.696938 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6224964-94b7-443c-abd6-4c101e4fa9fc" path="/var/lib/kubelet/pods/f6224964-94b7-443c-abd6-4c101e4fa9fc/volumes" Oct 03 15:17:48 crc kubenswrapper[4959]: I1003 15:17:48.051146 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxjgp"] Oct 03 15:17:48 crc kubenswrapper[4959]: I1003 15:17:48.068476 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxjgp"] Oct 03 15:17:48 crc kubenswrapper[4959]: I1003 15:17:48.690474 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerStarted","Data":"b65e5d21042ec86205f3b7217ee880c98352c3e10b4a0d17e2c01bbe9fb22b2b"} Oct 03 15:17:48 crc kubenswrapper[4959]: I1003 15:17:48.691092 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 15:17:48 crc kubenswrapper[4959]: I1003 15:17:48.715257 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.673124756 podStartE2EDuration="6.715239126s" podCreationTimestamp="2025-10-03 15:17:42 +0000 UTC" firstStartedPulling="2025-10-03 15:17:42.920315033 +0000 UTC m=+6432.123658450" lastFinishedPulling="2025-10-03 15:17:47.962429403 +0000 UTC m=+6437.165772820" observedRunningTime="2025-10-03 15:17:48.710575922 +0000 UTC m=+6437.913919349" watchObservedRunningTime="2025-10-03 15:17:48.715239126 +0000 UTC m=+6437.918582543" Oct 03 15:17:49 crc kubenswrapper[4959]: I1003 15:17:49.697946 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6d97872-2c24-4ee7-ae05-e599ef218ded" path="/var/lib/kubelet/pods/e6d97872-2c24-4ee7-ae05-e599ef218ded/volumes" Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.732322 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.736749 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.752401 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-w5d68"] Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.754104 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.765878 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-w5d68"] Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.793113 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 03 15:17:52 crc kubenswrapper[4959]: I1003 15:17:52.938621 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54hmd\" (UniqueName: \"kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd\") pod \"aodh-db-create-w5d68\" (UID: \"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7\") " pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.040447 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54hmd\" (UniqueName: \"kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd\") pod \"aodh-db-create-w5d68\" (UID: \"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7\") " pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.066014 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54hmd\" (UniqueName: \"kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd\") pod \"aodh-db-create-w5d68\" (UID: \"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7\") " pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.081930 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.406589 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-w5d68"] Oct 03 15:17:53 crc kubenswrapper[4959]: W1003 15:17:53.413616 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd412c3ef_82c9_46f5_a16c_8765c5f8d8c7.slice/crio-9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce WatchSource:0}: Error finding container 9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce: Status 404 returned error can't find the container with id 9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.804542 4959 generic.go:334] "Generic (PLEG): container finished" podID="d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" containerID="85c85355801d0989a8ca4ebe35ef7e490087c1512a3d43eb9b85a2079eec7da8" exitCode=0 Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.805877 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w5d68" event={"ID":"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7","Type":"ContainerDied","Data":"85c85355801d0989a8ca4ebe35ef7e490087c1512a3d43eb9b85a2079eec7da8"} Oct 03 15:17:53 crc kubenswrapper[4959]: I1003 15:17:53.805915 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w5d68" event={"ID":"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7","Type":"ContainerStarted","Data":"9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce"} Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.314337 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w5d68" Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.395141 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54hmd\" (UniqueName: \"kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd\") pod \"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7\" (UID: \"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7\") " Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.406440 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd" (OuterVolumeSpecName: "kube-api-access-54hmd") pod "d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" (UID: "d412c3ef-82c9-46f5-a16c-8765c5f8d8c7"). InnerVolumeSpecName "kube-api-access-54hmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.501990 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54hmd\" (UniqueName: \"kubernetes.io/projected/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7-kube-api-access-54hmd\") on node \"crc\" DevicePath \"\"" Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.836664 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-w5d68" event={"ID":"d412c3ef-82c9-46f5-a16c-8765c5f8d8c7","Type":"ContainerDied","Data":"9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce"} Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.836723 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9426b175f53d3736a7f9e6b2f73f7356e9cf09603639365f547bc6bdd94293ce" Oct 03 15:17:55 crc kubenswrapper[4959]: I1003 15:17:55.836728 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-w5d68" Oct 03 15:18:00 crc kubenswrapper[4959]: I1003 15:18:00.687216 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:18:00 crc kubenswrapper[4959]: E1003 15:18:00.688050 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:18:01 crc kubenswrapper[4959]: I1003 15:18:01.048740 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmnd8"] Oct 03 15:18:01 crc kubenswrapper[4959]: I1003 15:18:01.074613 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kmnd8"] Oct 03 15:18:01 crc kubenswrapper[4959]: I1003 15:18:01.706860 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d0c23d-fd4c-4027-9e06-7067e42c0c27" path="/var/lib/kubelet/pods/b6d0c23d-fd4c-4027-9e06-7067e42c0c27/volumes" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.896409 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-54f8-account-create-8kj78"] Oct 03 15:18:02 crc kubenswrapper[4959]: E1003 15:18:02.897602 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" containerName="mariadb-database-create" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.897625 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" containerName="mariadb-database-create" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.898050 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" containerName="mariadb-database-create" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.899316 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.901189 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.907142 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-54f8-account-create-8kj78"] Oct 03 15:18:02 crc kubenswrapper[4959]: I1003 15:18:02.990286 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm42m\" (UniqueName: \"kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m\") pod \"aodh-54f8-account-create-8kj78\" (UID: \"ea38a4f4-157e-442d-b56c-3026346f2849\") " pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:03 crc kubenswrapper[4959]: I1003 15:18:03.091559 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm42m\" (UniqueName: \"kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m\") pod \"aodh-54f8-account-create-8kj78\" (UID: \"ea38a4f4-157e-442d-b56c-3026346f2849\") " pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:03 crc kubenswrapper[4959]: I1003 15:18:03.113331 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm42m\" (UniqueName: \"kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m\") pod \"aodh-54f8-account-create-8kj78\" (UID: \"ea38a4f4-157e-442d-b56c-3026346f2849\") " pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:03 crc kubenswrapper[4959]: I1003 15:18:03.235147 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:03 crc kubenswrapper[4959]: I1003 15:18:03.753308 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-54f8-account-create-8kj78"] Oct 03 15:18:03 crc kubenswrapper[4959]: I1003 15:18:03.933800 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-54f8-account-create-8kj78" event={"ID":"ea38a4f4-157e-442d-b56c-3026346f2849","Type":"ContainerStarted","Data":"c854b00209e57dd45b96d17c627bc4994b6db65a664689a6d02c7c23b522f582"} Oct 03 15:18:04 crc kubenswrapper[4959]: I1003 15:18:04.947694 4959 generic.go:334] "Generic (PLEG): container finished" podID="ea38a4f4-157e-442d-b56c-3026346f2849" containerID="bc327590ada3a3adef141d2f334bde72c0e992fa75b81b7eba2a57acf85a81da" exitCode=0 Oct 03 15:18:04 crc kubenswrapper[4959]: I1003 15:18:04.947777 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-54f8-account-create-8kj78" event={"ID":"ea38a4f4-157e-442d-b56c-3026346f2849","Type":"ContainerDied","Data":"bc327590ada3a3adef141d2f334bde72c0e992fa75b81b7eba2a57acf85a81da"} Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.406218 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.466312 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm42m\" (UniqueName: \"kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m\") pod \"ea38a4f4-157e-442d-b56c-3026346f2849\" (UID: \"ea38a4f4-157e-442d-b56c-3026346f2849\") " Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.475422 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m" (OuterVolumeSpecName: "kube-api-access-fm42m") pod "ea38a4f4-157e-442d-b56c-3026346f2849" (UID: "ea38a4f4-157e-442d-b56c-3026346f2849"). InnerVolumeSpecName "kube-api-access-fm42m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.570180 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm42m\" (UniqueName: \"kubernetes.io/projected/ea38a4f4-157e-442d-b56c-3026346f2849-kube-api-access-fm42m\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.973750 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-54f8-account-create-8kj78" event={"ID":"ea38a4f4-157e-442d-b56c-3026346f2849","Type":"ContainerDied","Data":"c854b00209e57dd45b96d17c627bc4994b6db65a664689a6d02c7c23b522f582"} Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.973786 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c854b00209e57dd45b96d17c627bc4994b6db65a664689a6d02c7c23b522f582" Oct 03 15:18:06 crc kubenswrapper[4959]: I1003 15:18:06.973803 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-54f8-account-create-8kj78" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.323239 4959 scope.go:117] "RemoveContainer" containerID="21511c2b47aa72040fb41b48089d01a0aaab3c2ac55eff25fb5742ef7e6e051f" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.388390 4959 scope.go:117] "RemoveContainer" containerID="e8c96f23ea891ef8cc63d4f472fbf52821a7c50318ee5ed0071763b50e185bf6" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.405807 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-wqxdl"] Oct 03 15:18:08 crc kubenswrapper[4959]: E1003 15:18:08.406745 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea38a4f4-157e-442d-b56c-3026346f2849" containerName="mariadb-account-create" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.406796 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea38a4f4-157e-442d-b56c-3026346f2849" containerName="mariadb-account-create" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.407344 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea38a4f4-157e-442d-b56c-3026346f2849" containerName="mariadb-account-create" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.408766 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.410746 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-54c25" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.416000 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.416524 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.419554 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wqxdl"] Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.479264 4959 scope.go:117] "RemoveContainer" containerID="3eba02b4c8ab1dda9828a6b93401119c54a34d1ccb3ffb22475357bc91df9312" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.510722 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.510792 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.510849 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.510914 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzfsw\" (UniqueName: \"kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.512058 4959 scope.go:117] "RemoveContainer" containerID="4e02e273de6ad29814726865a5dc6c0f6a073c121670d02c2f8ef57325111b01" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.571987 4959 scope.go:117] "RemoveContainer" containerID="4a8746be50e608988b75314ad682cf7cbcadb2f4a3f65b66161430079f22f794" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.614845 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.614940 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.615025 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzfsw\" (UniqueName: \"kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.615250 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.626550 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.626935 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.627052 4959 scope.go:117] "RemoveContainer" containerID="6dd46a12ddbf3dba9ccebc3d17f454910073bad46a8a2e4082d2c192b7402413" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.628118 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.642127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzfsw\" (UniqueName: \"kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw\") pod \"aodh-db-sync-wqxdl\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.669739 4959 scope.go:117] "RemoveContainer" containerID="c3a5a97c89d7dbf3c92010ae6b739394db37ec5dfe1f4b205d4d3f0a4397277e" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.706168 4959 scope.go:117] "RemoveContainer" containerID="f86399ce1ead24b1b0f6d2633b1e9bc4f257535d260829ff27612e16f1254b07" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.798660 4959 scope.go:117] "RemoveContainer" containerID="eb1b282761a352975a2938dfa259624fd47ce90a79367510d6fec7b4dcb43e10" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.834464 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:08 crc kubenswrapper[4959]: I1003 15:18:08.841773 4959 scope.go:117] "RemoveContainer" containerID="0345bfd3ca139d031001ec4291e82685a5d5f5c411f3e8c535f85bce80318bfe" Oct 03 15:18:09 crc kubenswrapper[4959]: I1003 15:18:09.424990 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wqxdl"] Oct 03 15:18:10 crc kubenswrapper[4959]: I1003 15:18:10.068868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wqxdl" event={"ID":"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd","Type":"ContainerStarted","Data":"92d94267ba927698e9ce64980989fdd3c91c64f010a8eac637b0d96eb005613e"} Oct 03 15:18:12 crc kubenswrapper[4959]: I1003 15:18:12.377168 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 15:18:14 crc kubenswrapper[4959]: I1003 15:18:14.685778 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:18:14 crc kubenswrapper[4959]: E1003 15:18:14.686537 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:18:19 crc kubenswrapper[4959]: I1003 15:18:19.174781 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wqxdl" event={"ID":"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd","Type":"ContainerStarted","Data":"77743be64b83e19ee0316e2b9c15c3b54ad6cc15e7b78b53ee78c50bcaf80b6d"} Oct 03 15:18:19 crc kubenswrapper[4959]: I1003 15:18:19.204965 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-wqxdl" podStartSLOduration=2.277368714 podStartE2EDuration="11.20493907s" podCreationTimestamp="2025-10-03 15:18:08 +0000 UTC" firstStartedPulling="2025-10-03 15:18:09.430754607 +0000 UTC m=+6458.634098024" lastFinishedPulling="2025-10-03 15:18:18.358324963 +0000 UTC m=+6467.561668380" observedRunningTime="2025-10-03 15:18:19.19631787 +0000 UTC m=+6468.399661307" watchObservedRunningTime="2025-10-03 15:18:19.20493907 +0000 UTC m=+6468.408282507" Oct 03 15:18:21 crc kubenswrapper[4959]: I1003 15:18:21.198057 4959 generic.go:334] "Generic (PLEG): container finished" podID="0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" containerID="77743be64b83e19ee0316e2b9c15c3b54ad6cc15e7b78b53ee78c50bcaf80b6d" exitCode=0 Oct 03 15:18:21 crc kubenswrapper[4959]: I1003 15:18:21.198149 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wqxdl" event={"ID":"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd","Type":"ContainerDied","Data":"77743be64b83e19ee0316e2b9c15c3b54ad6cc15e7b78b53ee78c50bcaf80b6d"} Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.703551 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.842877 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data\") pod \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.843059 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle\") pod \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.843100 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzfsw\" (UniqueName: \"kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw\") pod \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.843211 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts\") pod \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\" (UID: \"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd\") " Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.848539 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts" (OuterVolumeSpecName: "scripts") pod "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" (UID: "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.849229 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw" (OuterVolumeSpecName: "kube-api-access-qzfsw") pod "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" (UID: "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd"). InnerVolumeSpecName "kube-api-access-qzfsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.875733 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data" (OuterVolumeSpecName: "config-data") pod "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" (UID: "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.884688 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" (UID: "0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.945613 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.945650 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.945665 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:22 crc kubenswrapper[4959]: I1003 15:18:22.945678 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzfsw\" (UniqueName: \"kubernetes.io/projected/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd-kube-api-access-qzfsw\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:23 crc kubenswrapper[4959]: I1003 15:18:23.229633 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wqxdl" event={"ID":"0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd","Type":"ContainerDied","Data":"92d94267ba927698e9ce64980989fdd3c91c64f010a8eac637b0d96eb005613e"} Oct 03 15:18:23 crc kubenswrapper[4959]: I1003 15:18:23.229679 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d94267ba927698e9ce64980989fdd3c91c64f010a8eac637b0d96eb005613e" Oct 03 15:18:23 crc kubenswrapper[4959]: I1003 15:18:23.229743 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wqxdl" Oct 03 15:18:25 crc kubenswrapper[4959]: I1003 15:18:25.686893 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:18:25 crc kubenswrapper[4959]: E1003 15:18:25.688237 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.758013 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 03 15:18:27 crc kubenswrapper[4959]: E1003 15:18:27.758740 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" containerName="aodh-db-sync" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.758756 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" containerName="aodh-db-sync" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.759001 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" containerName="aodh-db-sync" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.760871 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.763495 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.763846 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.776092 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-54c25" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.784233 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.855178 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-config-data\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.855337 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-scripts\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.855499 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.855612 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz84q\" (UniqueName: \"kubernetes.io/projected/614f87cc-6308-4228-aad1-a0eafb9f57a8-kube-api-access-zz84q\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.958158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.958277 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz84q\" (UniqueName: \"kubernetes.io/projected/614f87cc-6308-4228-aad1-a0eafb9f57a8-kube-api-access-zz84q\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.958443 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-config-data\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.958488 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-scripts\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.966127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.966251 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-config-data\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.985876 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz84q\" (UniqueName: \"kubernetes.io/projected/614f87cc-6308-4228-aad1-a0eafb9f57a8-kube-api-access-zz84q\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:27 crc kubenswrapper[4959]: I1003 15:18:27.989233 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/614f87cc-6308-4228-aad1-a0eafb9f57a8-scripts\") pod \"aodh-0\" (UID: \"614f87cc-6308-4228-aad1-a0eafb9f57a8\") " pod="openstack/aodh-0" Oct 03 15:18:28 crc kubenswrapper[4959]: I1003 15:18:28.094055 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 03 15:18:28 crc kubenswrapper[4959]: I1003 15:18:28.585600 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 03 15:18:29 crc kubenswrapper[4959]: I1003 15:18:29.294138 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"614f87cc-6308-4228-aad1-a0eafb9f57a8","Type":"ContainerStarted","Data":"906c942241e2e17219acf7f89e35740c59f99f9f3b40064cd2c06ca9c21c9ac8"} Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.305308 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"614f87cc-6308-4228-aad1-a0eafb9f57a8","Type":"ContainerStarted","Data":"6e2af1713d0353435ad8630a1d77711e0ca0f90062836856983444f6ba8244dc"} Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.446743 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.447151 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="proxy-httpd" containerID="cri-o://b65e5d21042ec86205f3b7217ee880c98352c3e10b4a0d17e2c01bbe9fb22b2b" gracePeriod=30 Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.447130 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="sg-core" containerID="cri-o://d27d367a5d3094456370467e4e26062d8c5819a48f9efaa0bdbf1bca5289c251" gracePeriod=30 Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.447132 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-notification-agent" containerID="cri-o://6067fa1dd65d9a134141290137b46517df72da5cf669202fd9908ecda943dbff" gracePeriod=30 Oct 03 15:18:30 crc kubenswrapper[4959]: I1003 15:18:30.448301 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-central-agent" containerID="cri-o://5a7e1166a0dcd3e69a8ad07b9b7a695b1625ff9044ea1bb8891a48de902d44b7" gracePeriod=30 Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.316813 4959 generic.go:334] "Generic (PLEG): container finished" podID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerID="b65e5d21042ec86205f3b7217ee880c98352c3e10b4a0d17e2c01bbe9fb22b2b" exitCode=0 Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.317094 4959 generic.go:334] "Generic (PLEG): container finished" podID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerID="d27d367a5d3094456370467e4e26062d8c5819a48f9efaa0bdbf1bca5289c251" exitCode=2 Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.317104 4959 generic.go:334] "Generic (PLEG): container finished" podID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerID="5a7e1166a0dcd3e69a8ad07b9b7a695b1625ff9044ea1bb8891a48de902d44b7" exitCode=0 Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.316871 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerDied","Data":"b65e5d21042ec86205f3b7217ee880c98352c3e10b4a0d17e2c01bbe9fb22b2b"} Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.317139 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerDied","Data":"d27d367a5d3094456370467e4e26062d8c5819a48f9efaa0bdbf1bca5289c251"} Oct 03 15:18:31 crc kubenswrapper[4959]: I1003 15:18:31.317152 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerDied","Data":"5a7e1166a0dcd3e69a8ad07b9b7a695b1625ff9044ea1bb8891a48de902d44b7"} Oct 03 15:18:32 crc kubenswrapper[4959]: I1003 15:18:32.331495 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"614f87cc-6308-4228-aad1-a0eafb9f57a8","Type":"ContainerStarted","Data":"9d72de15643e3190be42b96c9955b048ad6e58ba50fffa5283d344e44a11a407"} Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.343442 4959 generic.go:334] "Generic (PLEG): container finished" podID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerID="6067fa1dd65d9a134141290137b46517df72da5cf669202fd9908ecda943dbff" exitCode=0 Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.343654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerDied","Data":"6067fa1dd65d9a134141290137b46517df72da5cf669202fd9908ecda943dbff"} Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.594887 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.713913 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714296 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714331 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714409 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714572 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k7zf\" (UniqueName: \"kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714642 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts\") pod \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\" (UID: \"b1769ee7-ccfd-44d6-8bec-8e3a4697c358\") " Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.714768 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.715384 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.715762 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.715788 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.720090 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts" (OuterVolumeSpecName: "scripts") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.720523 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf" (OuterVolumeSpecName: "kube-api-access-6k7zf") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "kube-api-access-6k7zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.762457 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.806546 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.817494 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.817517 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.817527 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k7zf\" (UniqueName: \"kubernetes.io/projected/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-kube-api-access-6k7zf\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.817535 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.829111 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data" (OuterVolumeSpecName: "config-data") pod "b1769ee7-ccfd-44d6-8bec-8e3a4697c358" (UID: "b1769ee7-ccfd-44d6-8bec-8e3a4697c358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:18:33 crc kubenswrapper[4959]: I1003 15:18:33.919378 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1769ee7-ccfd-44d6-8bec-8e3a4697c358-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.354062 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.354046 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1769ee7-ccfd-44d6-8bec-8e3a4697c358","Type":"ContainerDied","Data":"f4afade29a440ccb599219b226ff29ff98337615fb9bb9aff2415b57e233e3c1"} Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.354126 4959 scope.go:117] "RemoveContainer" containerID="b65e5d21042ec86205f3b7217ee880c98352c3e10b4a0d17e2c01bbe9fb22b2b" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.356402 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"614f87cc-6308-4228-aad1-a0eafb9f57a8","Type":"ContainerStarted","Data":"2fb13fe460e5aba99ffa943c49b2cfddbf5dfd59b4479129c71f39f7820b5d66"} Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.378246 4959 scope.go:117] "RemoveContainer" containerID="d27d367a5d3094456370467e4e26062d8c5819a48f9efaa0bdbf1bca5289c251" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.438490 4959 scope.go:117] "RemoveContainer" containerID="6067fa1dd65d9a134141290137b46517df72da5cf669202fd9908ecda943dbff" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.464552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.475136 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.485045 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:34 crc kubenswrapper[4959]: E1003 15:18:34.485672 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-central-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.485695 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-central-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: E1003 15:18:34.485735 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-notification-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.485746 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-notification-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: E1003 15:18:34.485775 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="sg-core" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.485783 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="sg-core" Oct 03 15:18:34 crc kubenswrapper[4959]: E1003 15:18:34.485803 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="proxy-httpd" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.485811 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="proxy-httpd" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.486124 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-notification-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.486221 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="ceilometer-central-agent" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.486239 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="proxy-httpd" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.486255 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" containerName="sg-core" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.489313 4959 scope.go:117] "RemoveContainer" containerID="5a7e1166a0dcd3e69a8ad07b9b7a695b1625ff9044ea1bb8891a48de902d44b7" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.492947 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.496836 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.497724 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.498080 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.634692 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.634771 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlhfs\" (UniqueName: \"kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.634961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.635108 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.635212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.635327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.635659 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.738249 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.738313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlhfs\" (UniqueName: \"kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.738382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.738438 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.739877 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.739956 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.740151 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.740510 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.740708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.746338 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.756687 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.758109 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.758631 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlhfs\" (UniqueName: \"kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.759753 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " pod="openstack/ceilometer-0" Oct 03 15:18:34 crc kubenswrapper[4959]: I1003 15:18:34.842648 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:18:35 crc kubenswrapper[4959]: I1003 15:18:35.698313 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1769ee7-ccfd-44d6-8bec-8e3a4697c358" path="/var/lib/kubelet/pods/b1769ee7-ccfd-44d6-8bec-8e3a4697c358/volumes" Oct 03 15:18:35 crc kubenswrapper[4959]: I1003 15:18:35.731620 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:18:36 crc kubenswrapper[4959]: I1003 15:18:36.388456 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"614f87cc-6308-4228-aad1-a0eafb9f57a8","Type":"ContainerStarted","Data":"a51ccb3dc36f3f990ce366088294d783150800efcce0f819286a4a6efaa0895c"} Oct 03 15:18:36 crc kubenswrapper[4959]: I1003 15:18:36.390845 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerStarted","Data":"33fb9a90b8d3860e39ffa30f540df5fa03c0b2abebf2cd132ad0e60d59424737"} Oct 03 15:18:36 crc kubenswrapper[4959]: I1003 15:18:36.434430 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.738443767 podStartE2EDuration="9.4344008s" podCreationTimestamp="2025-10-03 15:18:27 +0000 UTC" firstStartedPulling="2025-10-03 15:18:28.599792781 +0000 UTC m=+6477.803136198" lastFinishedPulling="2025-10-03 15:18:35.295749814 +0000 UTC m=+6484.499093231" observedRunningTime="2025-10-03 15:18:36.405444421 +0000 UTC m=+6485.608787838" watchObservedRunningTime="2025-10-03 15:18:36.4344008 +0000 UTC m=+6485.637744227" Oct 03 15:18:37 crc kubenswrapper[4959]: I1003 15:18:37.409505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerStarted","Data":"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90"} Oct 03 15:18:37 crc kubenswrapper[4959]: I1003 15:18:37.409993 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerStarted","Data":"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112"} Oct 03 15:18:38 crc kubenswrapper[4959]: I1003 15:18:38.426285 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerStarted","Data":"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c"} Oct 03 15:18:40 crc kubenswrapper[4959]: I1003 15:18:40.686004 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:18:40 crc kubenswrapper[4959]: E1003 15:18:40.686775 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:18:41 crc kubenswrapper[4959]: I1003 15:18:41.462897 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerStarted","Data":"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849"} Oct 03 15:18:41 crc kubenswrapper[4959]: I1003 15:18:41.463144 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 15:18:41 crc kubenswrapper[4959]: I1003 15:18:41.492919 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.938400714 podStartE2EDuration="7.492897382s" podCreationTimestamp="2025-10-03 15:18:34 +0000 UTC" firstStartedPulling="2025-10-03 15:18:35.7422273 +0000 UTC m=+6484.945570717" lastFinishedPulling="2025-10-03 15:18:40.296723948 +0000 UTC m=+6489.500067385" observedRunningTime="2025-10-03 15:18:41.490711038 +0000 UTC m=+6490.694054495" watchObservedRunningTime="2025-10-03 15:18:41.492897382 +0000 UTC m=+6490.696240809" Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.390505 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-cpk86"] Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.399068 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cpk86" Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.403119 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-cpk86"] Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.532439 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh9l6\" (UniqueName: \"kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6\") pod \"manila-db-create-cpk86\" (UID: \"1910f143-5bca-4d35-a37f-82797c1bfdee\") " pod="openstack/manila-db-create-cpk86" Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.635043 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh9l6\" (UniqueName: \"kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6\") pod \"manila-db-create-cpk86\" (UID: \"1910f143-5bca-4d35-a37f-82797c1bfdee\") " pod="openstack/manila-db-create-cpk86" Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.654268 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh9l6\" (UniqueName: \"kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6\") pod \"manila-db-create-cpk86\" (UID: \"1910f143-5bca-4d35-a37f-82797c1bfdee\") " pod="openstack/manila-db-create-cpk86" Oct 03 15:18:42 crc kubenswrapper[4959]: I1003 15:18:42.734074 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cpk86" Oct 03 15:18:43 crc kubenswrapper[4959]: I1003 15:18:43.342842 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-cpk86"] Oct 03 15:18:43 crc kubenswrapper[4959]: W1003 15:18:43.352899 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1910f143_5bca_4d35_a37f_82797c1bfdee.slice/crio-9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee WatchSource:0}: Error finding container 9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee: Status 404 returned error can't find the container with id 9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee Oct 03 15:18:43 crc kubenswrapper[4959]: I1003 15:18:43.510345 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cpk86" event={"ID":"1910f143-5bca-4d35-a37f-82797c1bfdee","Type":"ContainerStarted","Data":"9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee"} Oct 03 15:18:44 crc kubenswrapper[4959]: I1003 15:18:44.523711 4959 generic.go:334] "Generic (PLEG): container finished" podID="1910f143-5bca-4d35-a37f-82797c1bfdee" containerID="460bdb5d9473ef24810f468a9b7a6aea0c4eb1862bdc0e90c1441f8796073180" exitCode=0 Oct 03 15:18:44 crc kubenswrapper[4959]: I1003 15:18:44.523782 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cpk86" event={"ID":"1910f143-5bca-4d35-a37f-82797c1bfdee","Type":"ContainerDied","Data":"460bdb5d9473ef24810f468a9b7a6aea0c4eb1862bdc0e90c1441f8796073180"} Oct 03 15:18:45 crc kubenswrapper[4959]: I1003 15:18:45.968237 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cpk86" Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.112179 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh9l6\" (UniqueName: \"kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6\") pod \"1910f143-5bca-4d35-a37f-82797c1bfdee\" (UID: \"1910f143-5bca-4d35-a37f-82797c1bfdee\") " Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.120488 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6" (OuterVolumeSpecName: "kube-api-access-kh9l6") pod "1910f143-5bca-4d35-a37f-82797c1bfdee" (UID: "1910f143-5bca-4d35-a37f-82797c1bfdee"). InnerVolumeSpecName "kube-api-access-kh9l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.215076 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh9l6\" (UniqueName: \"kubernetes.io/projected/1910f143-5bca-4d35-a37f-82797c1bfdee-kube-api-access-kh9l6\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.549448 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cpk86" event={"ID":"1910f143-5bca-4d35-a37f-82797c1bfdee","Type":"ContainerDied","Data":"9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee"} Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.549490 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bde69e66d25d26a67d5937d9ff2cc2320efa07dec2576b585addc79341d61ee" Oct 03 15:18:46 crc kubenswrapper[4959]: I1003 15:18:46.549531 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cpk86" Oct 03 15:18:47 crc kubenswrapper[4959]: I1003 15:18:47.062539 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-thcmv"] Oct 03 15:18:47 crc kubenswrapper[4959]: I1003 15:18:47.076097 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-thcmv"] Oct 03 15:18:47 crc kubenswrapper[4959]: I1003 15:18:47.701120 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f90f86b9-e22c-4f37-b44d-774879072a1c" path="/var/lib/kubelet/pods/f90f86b9-e22c-4f37-b44d-774879072a1c/volumes" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.529124 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-94da-account-create-m89k9"] Oct 03 15:18:52 crc kubenswrapper[4959]: E1003 15:18:52.530277 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1910f143-5bca-4d35-a37f-82797c1bfdee" containerName="mariadb-database-create" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.530294 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1910f143-5bca-4d35-a37f-82797c1bfdee" containerName="mariadb-database-create" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.530590 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1910f143-5bca-4d35-a37f-82797c1bfdee" containerName="mariadb-database-create" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.531497 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.534493 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.539181 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-94da-account-create-m89k9"] Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.658484 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqcf\" (UniqueName: \"kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf\") pod \"manila-94da-account-create-m89k9\" (UID: \"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e\") " pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.760816 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqcf\" (UniqueName: \"kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf\") pod \"manila-94da-account-create-m89k9\" (UID: \"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e\") " pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.786532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqcf\" (UniqueName: \"kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf\") pod \"manila-94da-account-create-m89k9\" (UID: \"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e\") " pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:52 crc kubenswrapper[4959]: I1003 15:18:52.853774 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:53 crc kubenswrapper[4959]: I1003 15:18:53.320019 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-94da-account-create-m89k9"] Oct 03 15:18:53 crc kubenswrapper[4959]: W1003 15:18:53.327071 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4e59dfb_dce5_47d2_a21e_dc87b6fc383e.slice/crio-c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b WatchSource:0}: Error finding container c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b: Status 404 returned error can't find the container with id c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b Oct 03 15:18:53 crc kubenswrapper[4959]: I1003 15:18:53.621385 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" containerID="c9629f524e34e5d89c359b8d21f736da0e2eae4c7298f5171deb2924ed252f53" exitCode=0 Oct 03 15:18:53 crc kubenswrapper[4959]: I1003 15:18:53.621478 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-94da-account-create-m89k9" event={"ID":"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e","Type":"ContainerDied","Data":"c9629f524e34e5d89c359b8d21f736da0e2eae4c7298f5171deb2924ed252f53"} Oct 03 15:18:53 crc kubenswrapper[4959]: I1003 15:18:53.621711 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-94da-account-create-m89k9" event={"ID":"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e","Type":"ContainerStarted","Data":"c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b"} Oct 03 15:18:54 crc kubenswrapper[4959]: I1003 15:18:54.685944 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:18:54 crc kubenswrapper[4959]: E1003 15:18:54.686692 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.074768 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.140243 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqcf\" (UniqueName: \"kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf\") pod \"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e\" (UID: \"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e\") " Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.147992 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf" (OuterVolumeSpecName: "kube-api-access-hxqcf") pod "f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" (UID: "f4e59dfb-dce5-47d2-a21e-dc87b6fc383e"). InnerVolumeSpecName "kube-api-access-hxqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.242606 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqcf\" (UniqueName: \"kubernetes.io/projected/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e-kube-api-access-hxqcf\") on node \"crc\" DevicePath \"\"" Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.642311 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-94da-account-create-m89k9" event={"ID":"f4e59dfb-dce5-47d2-a21e-dc87b6fc383e","Type":"ContainerDied","Data":"c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b"} Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.642356 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0e82b040e1f6921a1e6e9f2586e0c7a82efb11017d5a2d612108b90302c168b" Oct 03 15:18:55 crc kubenswrapper[4959]: I1003 15:18:55.642434 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-94da-account-create-m89k9" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.050454 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9d9b-account-create-bcdpk"] Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.065002 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9d9b-account-create-bcdpk"] Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.704096 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f8f9179-bec9-4149-8301-7ae6097d8153" path="/var/lib/kubelet/pods/1f8f9179-bec9-4149-8301-7ae6097d8153/volumes" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.868260 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-mfcn8"] Oct 03 15:18:57 crc kubenswrapper[4959]: E1003 15:18:57.868891 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" containerName="mariadb-account-create" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.868921 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" containerName="mariadb-account-create" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.869272 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" containerName="mariadb-account-create" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.870223 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.872685 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-hcsxp" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.872897 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.879935 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-mfcn8"] Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.901182 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.901304 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.901365 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:57 crc kubenswrapper[4959]: I1003 15:18:57.901542 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72fzb\" (UniqueName: \"kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.003393 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.003457 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.004562 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72fzb\" (UniqueName: \"kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.004706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.012778 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.013171 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.013426 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.025521 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72fzb\" (UniqueName: \"kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb\") pod \"manila-db-sync-mfcn8\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.194976 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-mfcn8" Oct 03 15:18:58 crc kubenswrapper[4959]: I1003 15:18:58.871072 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-mfcn8"] Oct 03 15:18:59 crc kubenswrapper[4959]: I1003 15:18:59.698099 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-mfcn8" event={"ID":"7153a88b-bb05-46b9-8df2-7dff0620e8ba","Type":"ContainerStarted","Data":"7c8121f8ab0c522abb139e4e996f9630f720486f6f35370a7379ce5a8b47b954"} Oct 03 15:19:04 crc kubenswrapper[4959]: I1003 15:19:04.029495 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-ztg4p"] Oct 03 15:19:04 crc kubenswrapper[4959]: I1003 15:19:04.050301 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-ztg4p"] Oct 03 15:19:04 crc kubenswrapper[4959]: I1003 15:19:04.742285 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-mfcn8" event={"ID":"7153a88b-bb05-46b9-8df2-7dff0620e8ba","Type":"ContainerStarted","Data":"cdb77196f0e9c931a9d3d986c1466a0e505df53d2458e5bb3e82a80a8e3e9d8c"} Oct 03 15:19:04 crc kubenswrapper[4959]: I1003 15:19:04.766810 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-mfcn8" podStartSLOduration=3.041335978 podStartE2EDuration="7.76679368s" podCreationTimestamp="2025-10-03 15:18:57 +0000 UTC" firstStartedPulling="2025-10-03 15:18:58.877790344 +0000 UTC m=+6508.081133761" lastFinishedPulling="2025-10-03 15:19:03.603248036 +0000 UTC m=+6512.806591463" observedRunningTime="2025-10-03 15:19:04.763379767 +0000 UTC m=+6513.966723184" watchObservedRunningTime="2025-10-03 15:19:04.76679368 +0000 UTC m=+6513.970137097" Oct 03 15:19:04 crc kubenswrapper[4959]: I1003 15:19:04.848533 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 15:19:05 crc kubenswrapper[4959]: I1003 15:19:05.703633 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aa49515-5afb-4d44-9aef-b5f8b0602773" path="/var/lib/kubelet/pods/1aa49515-5afb-4d44-9aef-b5f8b0602773/volumes" Oct 03 15:19:06 crc kubenswrapper[4959]: I1003 15:19:06.762639 4959 generic.go:334] "Generic (PLEG): container finished" podID="7153a88b-bb05-46b9-8df2-7dff0620e8ba" containerID="cdb77196f0e9c931a9d3d986c1466a0e505df53d2458e5bb3e82a80a8e3e9d8c" exitCode=0 Oct 03 15:19:06 crc kubenswrapper[4959]: I1003 15:19:06.762923 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-mfcn8" event={"ID":"7153a88b-bb05-46b9-8df2-7dff0620e8ba","Type":"ContainerDied","Data":"cdb77196f0e9c931a9d3d986c1466a0e505df53d2458e5bb3e82a80a8e3e9d8c"} Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.293562 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-mfcn8" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.449570 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data\") pod \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.449906 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72fzb\" (UniqueName: \"kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb\") pod \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.449967 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data\") pod \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.450064 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle\") pod \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\" (UID: \"7153a88b-bb05-46b9-8df2-7dff0620e8ba\") " Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.455500 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "7153a88b-bb05-46b9-8df2-7dff0620e8ba" (UID: "7153a88b-bb05-46b9-8df2-7dff0620e8ba"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.455642 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb" (OuterVolumeSpecName: "kube-api-access-72fzb") pod "7153a88b-bb05-46b9-8df2-7dff0620e8ba" (UID: "7153a88b-bb05-46b9-8df2-7dff0620e8ba"). InnerVolumeSpecName "kube-api-access-72fzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.457671 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data" (OuterVolumeSpecName: "config-data") pod "7153a88b-bb05-46b9-8df2-7dff0620e8ba" (UID: "7153a88b-bb05-46b9-8df2-7dff0620e8ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.500299 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7153a88b-bb05-46b9-8df2-7dff0620e8ba" (UID: "7153a88b-bb05-46b9-8df2-7dff0620e8ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.552306 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.552348 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72fzb\" (UniqueName: \"kubernetes.io/projected/7153a88b-bb05-46b9-8df2-7dff0620e8ba-kube-api-access-72fzb\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.552361 4959 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.552373 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7153a88b-bb05-46b9-8df2-7dff0620e8ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.785385 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-mfcn8" event={"ID":"7153a88b-bb05-46b9-8df2-7dff0620e8ba","Type":"ContainerDied","Data":"7c8121f8ab0c522abb139e4e996f9630f720486f6f35370a7379ce5a8b47b954"} Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.785428 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c8121f8ab0c522abb139e4e996f9630f720486f6f35370a7379ce5a8b47b954" Oct 03 15:19:08 crc kubenswrapper[4959]: I1003 15:19:08.785438 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-mfcn8" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.163576 4959 scope.go:117] "RemoveContainer" containerID="a8a8ccb2c084ae66d812cc1ec4033b06a20329b0cf13dbecaf9788172090a69d" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.248464 4959 scope.go:117] "RemoveContainer" containerID="99a8906059e98f47b01e40bc5bc64c175edb24eef139f2c4a27eee3bdaa64f90" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.249957 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:19:09 crc kubenswrapper[4959]: E1003 15:19:09.250502 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7153a88b-bb05-46b9-8df2-7dff0620e8ba" containerName="manila-db-sync" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.250515 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7153a88b-bb05-46b9-8df2-7dff0620e8ba" containerName="manila-db-sync" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.250723 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7153a88b-bb05-46b9-8df2-7dff0620e8ba" containerName="manila-db-sync" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.251870 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.266955 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.269112 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.272548 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.272849 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.273006 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-hcsxp" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.273097 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.288301 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.288455 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.288492 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbh7s\" (UniqueName: \"kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.288519 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.288614 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.289835 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.314643 4959 scope.go:117] "RemoveContainer" containerID="b0f425f4be4223ff24761ccf089044aa8b86b954982ffe4ca020bd410a614b61" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.320014 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.322147 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.328686 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.332374 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.347149 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.371690 4959 scope.go:117] "RemoveContainer" containerID="7ab40e3fdd0e277abc0eb32fcca5c08831d7140587aa96478902a76e532be5fb" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.390352 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.390441 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.390466 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbh7s\" (UniqueName: \"kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.390489 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.390536 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.391434 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.392824 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.397335 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.418484 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.450355 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbh7s\" (UniqueName: \"kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s\") pod \"dnsmasq-dns-55586cc989-q6ngx\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.517912 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-scripts\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.517963 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9qj2\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-kube-api-access-j9qj2\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.517997 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518061 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518106 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-scripts\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518239 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518305 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518327 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518346 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwx2j\" (UniqueName: \"kubernetes.io/projected/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-kube-api-access-mwx2j\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518495 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518532 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-ceph\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518557 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.518589 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.527834 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.529554 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.543012 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620835 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620888 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-ceph\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620914 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620937 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620959 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-scripts\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620975 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9qj2\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-kube-api-access-j9qj2\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.620998 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621050 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a804906c-740a-4406-bdd5-44ab84050ee8-etc-machine-id\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621074 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621100 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621123 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-scripts\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621182 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhz8k\" (UniqueName: \"kubernetes.io/projected/a804906c-740a-4406-bdd5-44ab84050ee8-kube-api-access-hhz8k\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621230 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621250 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621306 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621323 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwx2j\" (UniqueName: \"kubernetes.io/projected/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-kube-api-access-mwx2j\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621372 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621403 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-scripts\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621421 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data-custom\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.621438 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a804906c-740a-4406-bdd5-44ab84050ee8-logs\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.627644 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.627720 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.628716 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.639218 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2e142326-f734-4066-9b6c-d7a587851ca5-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.645954 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.648983 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.652370 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.652876 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-scripts\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.688034 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:19:09 crc kubenswrapper[4959]: E1003 15:19:09.688364 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.691809 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9qj2\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-kube-api-access-j9qj2\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.692301 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-scripts\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.698101 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2e142326-f734-4066-9b6c-d7a587851ca5-ceph\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.699329 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.701456 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.702936 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e142326-f734-4066-9b6c-d7a587851ca5-config-data\") pod \"manila-share-share1-0\" (UID: \"2e142326-f734-4066-9b6c-d7a587851ca5\") " pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.710964 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-config-data\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.720709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.739936 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwx2j\" (UniqueName: \"kubernetes.io/projected/ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910-kube-api-access-mwx2j\") pod \"manila-scheduler-0\" (UID: \"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910\") " pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.743621 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data-custom\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.743767 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a804906c-740a-4406-bdd5-44ab84050ee8-logs\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.744405 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a804906c-740a-4406-bdd5-44ab84050ee8-etc-machine-id\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.744983 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.745632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhz8k\" (UniqueName: \"kubernetes.io/projected/a804906c-740a-4406-bdd5-44ab84050ee8-kube-api-access-hhz8k\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.745676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.746147 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-scripts\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.748143 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a804906c-740a-4406-bdd5-44ab84050ee8-etc-machine-id\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.770716 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.771402 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a804906c-740a-4406-bdd5-44ab84050ee8-logs\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.776305 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.777890 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-scripts\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.781770 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-config-data-custom\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.791364 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhz8k\" (UniqueName: \"kubernetes.io/projected/a804906c-740a-4406-bdd5-44ab84050ee8-kube-api-access-hhz8k\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.792698 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a804906c-740a-4406-bdd5-44ab84050ee8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"a804906c-740a-4406-bdd5-44ab84050ee8\") " pod="openstack/manila-api-0" Oct 03 15:19:09 crc kubenswrapper[4959]: I1003 15:19:09.914490 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.460482 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:19:10 crc kubenswrapper[4959]: W1003 15:19:10.500087 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6447897_6e23_41ff_a9a4_7c589fdb8d44.slice/crio-0ad4bd39aae5835771eaf34349083c3511ce0c7d0757230b03e9bfd373ddd01e WatchSource:0}: Error finding container 0ad4bd39aae5835771eaf34349083c3511ce0c7d0757230b03e9bfd373ddd01e: Status 404 returned error can't find the container with id 0ad4bd39aae5835771eaf34349083c3511ce0c7d0757230b03e9bfd373ddd01e Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.603674 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.729768 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.836924 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910","Type":"ContainerStarted","Data":"828edf9de485470b8b6cccb447155d931bff8f44d7acafc030ecaf3baf4ff19e"} Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.838658 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerStarted","Data":"94453bf503d05b3330efae4aa0368a92997bf88379d0c2a44c3fc0200860ea45"} Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.838683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerStarted","Data":"0ad4bd39aae5835771eaf34349083c3511ce0c7d0757230b03e9bfd373ddd01e"} Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.842445 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2e142326-f734-4066-9b6c-d7a587851ca5","Type":"ContainerStarted","Data":"4eae686af70789e4230d435b40461a11851f38b3caa9abe70167e775feb2922a"} Oct 03 15:19:10 crc kubenswrapper[4959]: W1003 15:19:10.924364 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda804906c_740a_4406_bdd5_44ab84050ee8.slice/crio-48ecace841ec583e3ca999a79aa53de33ac514d65671cc77d1affb5d7997ea24 WatchSource:0}: Error finding container 48ecace841ec583e3ca999a79aa53de33ac514d65671cc77d1affb5d7997ea24: Status 404 returned error can't find the container with id 48ecace841ec583e3ca999a79aa53de33ac514d65671cc77d1affb5d7997ea24 Oct 03 15:19:10 crc kubenswrapper[4959]: I1003 15:19:10.924775 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 15:19:11 crc kubenswrapper[4959]: I1003 15:19:11.863056 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"a804906c-740a-4406-bdd5-44ab84050ee8","Type":"ContainerStarted","Data":"e9af749abd7c1bc89ac61a16d1ce6d7012c16f510a8872190156c80b0eb9b167"} Oct 03 15:19:11 crc kubenswrapper[4959]: I1003 15:19:11.863678 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"a804906c-740a-4406-bdd5-44ab84050ee8","Type":"ContainerStarted","Data":"48ecace841ec583e3ca999a79aa53de33ac514d65671cc77d1affb5d7997ea24"} Oct 03 15:19:11 crc kubenswrapper[4959]: I1003 15:19:11.867783 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerID="94453bf503d05b3330efae4aa0368a92997bf88379d0c2a44c3fc0200860ea45" exitCode=0 Oct 03 15:19:11 crc kubenswrapper[4959]: I1003 15:19:11.867829 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerDied","Data":"94453bf503d05b3330efae4aa0368a92997bf88379d0c2a44c3fc0200860ea45"} Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.888173 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910","Type":"ContainerStarted","Data":"bd799885371a5b85af405f0c1b8bef2bcb17d6d82f2a2d5db2d139cc5739aa3e"} Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.888708 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910","Type":"ContainerStarted","Data":"a5c6e02ca4aeabfc1aa15d981a27e112ec4e42d8a8cd90f22bfcfed76a6bc59e"} Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.898999 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerStarted","Data":"473beaff3bab8d1997baebfc4ba75658031b86e24da84ee1ead4015132585ec9"} Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.899936 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.920324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"a804906c-740a-4406-bdd5-44ab84050ee8","Type":"ContainerStarted","Data":"423c4fe57934ab801f4b36da557427e8d0fa48957e5f343427d370dcd7ed8efb"} Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.920964 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.922495 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.089619634 podStartE2EDuration="3.922486236s" podCreationTimestamp="2025-10-03 15:19:09 +0000 UTC" firstStartedPulling="2025-10-03 15:19:10.62452832 +0000 UTC m=+6519.827871737" lastFinishedPulling="2025-10-03 15:19:11.457394922 +0000 UTC m=+6520.660738339" observedRunningTime="2025-10-03 15:19:12.916668383 +0000 UTC m=+6522.120011800" watchObservedRunningTime="2025-10-03 15:19:12.922486236 +0000 UTC m=+6522.125829643" Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.959726 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" podStartSLOduration=3.959701767 podStartE2EDuration="3.959701767s" podCreationTimestamp="2025-10-03 15:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:19:12.933506505 +0000 UTC m=+6522.136849922" watchObservedRunningTime="2025-10-03 15:19:12.959701767 +0000 UTC m=+6522.163045184" Oct 03 15:19:12 crc kubenswrapper[4959]: I1003 15:19:12.963790 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.963776946 podStartE2EDuration="3.963776946s" podCreationTimestamp="2025-10-03 15:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:19:12.952725216 +0000 UTC m=+6522.156068623" watchObservedRunningTime="2025-10-03 15:19:12.963776946 +0000 UTC m=+6522.167120363" Oct 03 15:19:18 crc kubenswrapper[4959]: I1003 15:19:18.984683 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2e142326-f734-4066-9b6c-d7a587851ca5","Type":"ContainerStarted","Data":"04ac417d88e122a907cdf23da2339c35ca7ffd3b540b3aa738908df0dfcbf9a4"} Oct 03 15:19:18 crc kubenswrapper[4959]: I1003 15:19:18.985295 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2e142326-f734-4066-9b6c-d7a587851ca5","Type":"ContainerStarted","Data":"b93952923507aa585d1fe3ea23a0d2c3512b0854019c9c81280a5b7bb91a487e"} Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.023098 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.727236376 podStartE2EDuration="10.023078369s" podCreationTimestamp="2025-10-03 15:19:09 +0000 UTC" firstStartedPulling="2025-10-03 15:19:10.733401205 +0000 UTC m=+6519.936744622" lastFinishedPulling="2025-10-03 15:19:18.029243178 +0000 UTC m=+6527.232586615" observedRunningTime="2025-10-03 15:19:19.015125314 +0000 UTC m=+6528.218468781" watchObservedRunningTime="2025-10-03 15:19:19.023078369 +0000 UTC m=+6528.226421796" Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.654434 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.725930 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.739848 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.740102 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="dnsmasq-dns" containerID="cri-o://26a1934fc2088528fd226682b170dd222603ac5c4842aa13d2c59c844fc2ff15" gracePeriod=10 Oct 03 15:19:19 crc kubenswrapper[4959]: I1003 15:19:19.771889 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.000744 4959 generic.go:334] "Generic (PLEG): container finished" podID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerID="26a1934fc2088528fd226682b170dd222603ac5c4842aa13d2c59c844fc2ff15" exitCode=0 Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.000833 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" event={"ID":"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece","Type":"ContainerDied","Data":"26a1934fc2088528fd226682b170dd222603ac5c4842aa13d2c59c844fc2ff15"} Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.285779 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.346456 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s74t4\" (UniqueName: \"kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4\") pod \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.346576 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb\") pod \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.346601 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc\") pod \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.346682 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config\") pod \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.346878 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb\") pod \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\" (UID: \"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece\") " Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.369165 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4" (OuterVolumeSpecName: "kube-api-access-s74t4") pod "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" (UID: "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece"). InnerVolumeSpecName "kube-api-access-s74t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.416981 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config" (OuterVolumeSpecName: "config") pod "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" (UID: "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.417018 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" (UID: "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.420182 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" (UID: "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.434433 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" (UID: "f4e48b9f-e4d8-49d7-ab1c-e58a62598ece"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.449520 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.449551 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s74t4\" (UniqueName: \"kubernetes.io/projected/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-kube-api-access-s74t4\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.449563 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.449572 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:20 crc kubenswrapper[4959]: I1003 15:19:20.449581 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.011808 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.012512 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dc68b6c7-wb7jn" event={"ID":"f4e48b9f-e4d8-49d7-ab1c-e58a62598ece","Type":"ContainerDied","Data":"121dfff4b40b9d4b73c9273c05389190797f8ce0460190df6344ea105f40ab1d"} Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.012562 4959 scope.go:117] "RemoveContainer" containerID="26a1934fc2088528fd226682b170dd222603ac5c4842aa13d2c59c844fc2ff15" Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.043712 4959 scope.go:117] "RemoveContainer" containerID="ebd5639345efb2019d32c5f977f74c267a945b720c6a74ce75376c696f03e792" Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.052609 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.064677 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dc68b6c7-wb7jn"] Oct 03 15:19:21 crc kubenswrapper[4959]: I1003 15:19:21.699945 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" path="/var/lib/kubelet/pods/f4e48b9f-e4d8-49d7-ab1c-e58a62598ece/volumes" Oct 03 15:19:22 crc kubenswrapper[4959]: I1003 15:19:22.809454 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:22 crc kubenswrapper[4959]: I1003 15:19:22.809700 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-central-agent" containerID="cri-o://d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112" gracePeriod=30 Oct 03 15:19:22 crc kubenswrapper[4959]: I1003 15:19:22.810107 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="proxy-httpd" containerID="cri-o://866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849" gracePeriod=30 Oct 03 15:19:22 crc kubenswrapper[4959]: I1003 15:19:22.810149 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="sg-core" containerID="cri-o://fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c" gracePeriod=30 Oct 03 15:19:22 crc kubenswrapper[4959]: I1003 15:19:22.810185 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-notification-agent" containerID="cri-o://fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90" gracePeriod=30 Oct 03 15:19:23 crc kubenswrapper[4959]: I1003 15:19:23.039701 4959 generic.go:334] "Generic (PLEG): container finished" podID="a32af7a5-e329-4608-8209-67d03fcd7215" containerID="866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849" exitCode=0 Oct 03 15:19:23 crc kubenswrapper[4959]: I1003 15:19:23.040034 4959 generic.go:334] "Generic (PLEG): container finished" podID="a32af7a5-e329-4608-8209-67d03fcd7215" containerID="fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c" exitCode=2 Oct 03 15:19:23 crc kubenswrapper[4959]: I1003 15:19:23.040062 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerDied","Data":"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849"} Oct 03 15:19:23 crc kubenswrapper[4959]: I1003 15:19:23.040093 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerDied","Data":"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c"} Oct 03 15:19:23 crc kubenswrapper[4959]: E1003 15:19:23.096922 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32af7a5_e329_4608_8209_67d03fcd7215.slice/crio-fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32af7a5_e329_4608_8209_67d03fcd7215.slice/crio-conmon-fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32af7a5_e329_4608_8209_67d03fcd7215.slice/crio-conmon-866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32af7a5_e329_4608_8209_67d03fcd7215.slice/crio-866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849.scope\": RecentStats: unable to find data in memory cache]" Oct 03 15:19:24 crc kubenswrapper[4959]: I1003 15:19:24.088052 4959 generic.go:334] "Generic (PLEG): container finished" podID="a32af7a5-e329-4608-8209-67d03fcd7215" containerID="d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112" exitCode=0 Oct 03 15:19:24 crc kubenswrapper[4959]: I1003 15:19:24.088343 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerDied","Data":"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112"} Oct 03 15:19:24 crc kubenswrapper[4959]: I1003 15:19:24.685239 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:19:24 crc kubenswrapper[4959]: E1003 15:19:24.685493 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.797279 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880084 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880129 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880200 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880222 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880259 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlhfs\" (UniqueName: \"kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880441 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.880482 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts\") pod \"a32af7a5-e329-4608-8209-67d03fcd7215\" (UID: \"a32af7a5-e329-4608-8209-67d03fcd7215\") " Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.881990 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.882458 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.889301 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts" (OuterVolumeSpecName: "scripts") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.891186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs" (OuterVolumeSpecName: "kube-api-access-zlhfs") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "kube-api-access-zlhfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.934126 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.975467 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.982842 4959 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.986272 4959 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.986394 4959 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.986518 4959 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a32af7a5-e329-4608-8209-67d03fcd7215-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.986615 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.986696 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlhfs\" (UniqueName: \"kubernetes.io/projected/a32af7a5-e329-4608-8209-67d03fcd7215-kube-api-access-zlhfs\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:25 crc kubenswrapper[4959]: I1003 15:19:25.993648 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data" (OuterVolumeSpecName: "config-data") pod "a32af7a5-e329-4608-8209-67d03fcd7215" (UID: "a32af7a5-e329-4608-8209-67d03fcd7215"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.089546 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32af7a5-e329-4608-8209-67d03fcd7215-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.109123 4959 generic.go:334] "Generic (PLEG): container finished" podID="a32af7a5-e329-4608-8209-67d03fcd7215" containerID="fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90" exitCode=0 Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.109163 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerDied","Data":"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90"} Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.109205 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a32af7a5-e329-4608-8209-67d03fcd7215","Type":"ContainerDied","Data":"33fb9a90b8d3860e39ffa30f540df5fa03c0b2abebf2cd132ad0e60d59424737"} Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.109226 4959 scope.go:117] "RemoveContainer" containerID="866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.109277 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.130752 4959 scope.go:117] "RemoveContainer" containerID="fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.141322 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.150600 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.155348 4959 scope.go:117] "RemoveContainer" containerID="fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.168706 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169148 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-notification-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169163 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-notification-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169183 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="dnsmasq-dns" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169234 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="dnsmasq-dns" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169254 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="proxy-httpd" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169260 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="proxy-httpd" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169272 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="sg-core" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169278 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="sg-core" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169297 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-central-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169303 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-central-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.169315 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="init" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169321 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="init" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169518 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-notification-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169529 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="ceilometer-central-agent" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169542 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4e48b9f-e4d8-49d7-ab1c-e58a62598ece" containerName="dnsmasq-dns" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169557 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="proxy-httpd" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.169570 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" containerName="sg-core" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.171384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.173858 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.174235 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.178353 4959 scope.go:117] "RemoveContainer" containerID="d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.181155 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.209066 4959 scope.go:117] "RemoveContainer" containerID="866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.209577 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849\": container with ID starting with 866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849 not found: ID does not exist" containerID="866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.209649 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849"} err="failed to get container status \"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849\": rpc error: code = NotFound desc = could not find container \"866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849\": container with ID starting with 866897d59fdeb748b95e785cc2683cfacc1957d331a3586273e3caededcf6849 not found: ID does not exist" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.209683 4959 scope.go:117] "RemoveContainer" containerID="fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.210362 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c\": container with ID starting with fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c not found: ID does not exist" containerID="fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.210401 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c"} err="failed to get container status \"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c\": rpc error: code = NotFound desc = could not find container \"fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c\": container with ID starting with fa4c175aed2cfc4b0ceaeabfde968ed048129aa0d367bb0aeb346606bd7a8a3c not found: ID does not exist" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.210426 4959 scope.go:117] "RemoveContainer" containerID="fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.210705 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90\": container with ID starting with fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90 not found: ID does not exist" containerID="fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.210743 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90"} err="failed to get container status \"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90\": rpc error: code = NotFound desc = could not find container \"fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90\": container with ID starting with fc50211711d7fbe72f042a538dc16cd8dd7ed8110c2256acfd34ee383f189f90 not found: ID does not exist" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.210768 4959 scope.go:117] "RemoveContainer" containerID="d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112" Oct 03 15:19:26 crc kubenswrapper[4959]: E1003 15:19:26.211130 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112\": container with ID starting with d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112 not found: ID does not exist" containerID="d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.211154 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112"} err="failed to get container status \"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112\": rpc error: code = NotFound desc = could not find container \"d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112\": container with ID starting with d6930dbdf85ca5bedb8768a0a4e3e320845ad52b4596a61a5829d5d93f2ee112 not found: ID does not exist" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293451 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-run-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293510 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293609 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t45x\" (UniqueName: \"kubernetes.io/projected/5922a823-0599-4108-a7ca-d7eb68de6c2d-kube-api-access-7t45x\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293780 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-config-data\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293902 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.293937 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-log-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.294010 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-scripts\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395340 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395386 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-log-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395431 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-scripts\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-run-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395557 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t45x\" (UniqueName: \"kubernetes.io/projected/5922a823-0599-4108-a7ca-d7eb68de6c2d-kube-api-access-7t45x\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395580 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-config-data\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.395839 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-log-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.396447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5922a823-0599-4108-a7ca-d7eb68de6c2d-run-httpd\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.399422 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-config-data\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.399545 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.399905 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.402883 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5922a823-0599-4108-a7ca-d7eb68de6c2d-scripts\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.413444 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t45x\" (UniqueName: \"kubernetes.io/projected/5922a823-0599-4108-a7ca-d7eb68de6c2d-kube-api-access-7t45x\") pod \"ceilometer-0\" (UID: \"5922a823-0599-4108-a7ca-d7eb68de6c2d\") " pod="openstack/ceilometer-0" Oct 03 15:19:26 crc kubenswrapper[4959]: I1003 15:19:26.499478 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 15:19:27 crc kubenswrapper[4959]: I1003 15:19:27.058844 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 15:19:27 crc kubenswrapper[4959]: W1003 15:19:27.064462 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5922a823_0599_4108_a7ca_d7eb68de6c2d.slice/crio-8606962a4905558bab2a85ac921d0f828990e835e86a25e9c2765e782d459504 WatchSource:0}: Error finding container 8606962a4905558bab2a85ac921d0f828990e835e86a25e9c2765e782d459504: Status 404 returned error can't find the container with id 8606962a4905558bab2a85ac921d0f828990e835e86a25e9c2765e782d459504 Oct 03 15:19:27 crc kubenswrapper[4959]: I1003 15:19:27.123758 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"8606962a4905558bab2a85ac921d0f828990e835e86a25e9c2765e782d459504"} Oct 03 15:19:27 crc kubenswrapper[4959]: I1003 15:19:27.697387 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32af7a5-e329-4608-8209-67d03fcd7215" path="/var/lib/kubelet/pods/a32af7a5-e329-4608-8209-67d03fcd7215/volumes" Oct 03 15:19:28 crc kubenswrapper[4959]: I1003 15:19:28.140389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"8a8b567a92b733c71dbbe9b9b83175fe00dbd21de71a2cfc19d1cbdab50d31cf"} Oct 03 15:19:29 crc kubenswrapper[4959]: I1003 15:19:29.154855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"29097b82b5456dc9ca66762a31428a87e3c8f5c6856b29170584e4d9f8be1150"} Oct 03 15:19:30 crc kubenswrapper[4959]: I1003 15:19:30.179558 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"abec9c225efc507e813964e5c954a29f07031f5ce3eb8fcc415399cb3fbc3079"} Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.190901 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"0e06ab60408ea5b2908f5fb18c14d2b6ba4f09c23b512e9ea7377a2d9994e19c"} Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.191265 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.213843 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.372062836 podStartE2EDuration="5.213828372s" podCreationTimestamp="2025-10-03 15:19:26 +0000 UTC" firstStartedPulling="2025-10-03 15:19:27.071587593 +0000 UTC m=+6536.274931010" lastFinishedPulling="2025-10-03 15:19:30.913353129 +0000 UTC m=+6540.116696546" observedRunningTime="2025-10-03 15:19:31.210820299 +0000 UTC m=+6540.414163716" watchObservedRunningTime="2025-10-03 15:19:31.213828372 +0000 UTC m=+6540.417171789" Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.224955 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.237676 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 03 15:19:31 crc kubenswrapper[4959]: I1003 15:19:31.410368 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 03 15:19:39 crc kubenswrapper[4959]: I1003 15:19:39.685720 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:19:39 crc kubenswrapper[4959]: E1003 15:19:39.686659 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:19:52 crc kubenswrapper[4959]: I1003 15:19:52.685403 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:19:52 crc kubenswrapper[4959]: E1003 15:19:52.686183 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:19:56 crc kubenswrapper[4959]: I1003 15:19:56.509106 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 15:20:07 crc kubenswrapper[4959]: I1003 15:20:07.686004 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:20:07 crc kubenswrapper[4959]: E1003 15:20:07.686885 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.321271 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.323729 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.328985 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.350137 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451505 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451631 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh242\" (UniqueName: \"kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451693 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451822 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.451984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.554156 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.554352 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.554432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.555241 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.555314 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.555327 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.555625 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh242\" (UniqueName: \"kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.555706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.556330 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.556382 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.557063 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.581092 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh242\" (UniqueName: \"kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242\") pod \"dnsmasq-dns-7b54c866bc-qzbbr\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:17 crc kubenswrapper[4959]: I1003 15:20:17.650527 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:18 crc kubenswrapper[4959]: I1003 15:20:18.153508 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:18 crc kubenswrapper[4959]: I1003 15:20:18.762415 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerID="eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6" exitCode=0 Oct 03 15:20:18 crc kubenswrapper[4959]: I1003 15:20:18.762503 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" event={"ID":"3a256b35-130b-41bd-bdcc-24fc141c08cd","Type":"ContainerDied","Data":"eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6"} Oct 03 15:20:18 crc kubenswrapper[4959]: I1003 15:20:18.762932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" event={"ID":"3a256b35-130b-41bd-bdcc-24fc141c08cd","Type":"ContainerStarted","Data":"b70bb9dcb1b793929d4236dd805781e2b466dd8c41b5595089c59b858529faed"} Oct 03 15:20:19 crc kubenswrapper[4959]: I1003 15:20:19.787843 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" event={"ID":"3a256b35-130b-41bd-bdcc-24fc141c08cd","Type":"ContainerStarted","Data":"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504"} Oct 03 15:20:19 crc kubenswrapper[4959]: I1003 15:20:19.815362 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" podStartSLOduration=2.81530542 podStartE2EDuration="2.81530542s" podCreationTimestamp="2025-10-03 15:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:20:19.809165959 +0000 UTC m=+6589.012509416" watchObservedRunningTime="2025-10-03 15:20:19.81530542 +0000 UTC m=+6589.018648827" Oct 03 15:20:20 crc kubenswrapper[4959]: I1003 15:20:20.685920 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:20:20 crc kubenswrapper[4959]: E1003 15:20:20.686489 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:20:20 crc kubenswrapper[4959]: I1003 15:20:20.796680 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.652109 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.725284 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.725529 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="dnsmasq-dns" containerID="cri-o://473beaff3bab8d1997baebfc4ba75658031b86e24da84ee1ead4015132585ec9" gracePeriod=10 Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.872162 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d6cd869d9-km27m"] Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.874158 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.907845 4959 generic.go:334] "Generic (PLEG): container finished" podID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerID="473beaff3bab8d1997baebfc4ba75658031b86e24da84ee1ead4015132585ec9" exitCode=0 Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.907921 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerDied","Data":"473beaff3bab8d1997baebfc4ba75658031b86e24da84ee1ead4015132585ec9"} Oct 03 15:20:27 crc kubenswrapper[4959]: I1003 15:20:27.913332 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6cd869d9-km27m"] Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.027934 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-config\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.028086 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4h4j\" (UniqueName: \"kubernetes.io/projected/30fd93eb-4baf-471e-9dd3-28a623a2a954-kube-api-access-v4h4j\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.028237 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-openstack-cell1\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.028357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-nb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.028395 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-dns-svc\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.028645 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-sb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.131419 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-nb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.131715 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-dns-svc\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.131858 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-sb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.131989 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-config\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.132057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4h4j\" (UniqueName: \"kubernetes.io/projected/30fd93eb-4baf-471e-9dd3-28a623a2a954-kube-api-access-v4h4j\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.132095 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-openstack-cell1\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.132803 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-sb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.132838 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-ovsdbserver-nb\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.132912 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-dns-svc\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.133044 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-openstack-cell1\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.133636 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30fd93eb-4baf-471e-9dd3-28a623a2a954-config\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.172294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4h4j\" (UniqueName: \"kubernetes.io/projected/30fd93eb-4baf-471e-9dd3-28a623a2a954-kube-api-access-v4h4j\") pod \"dnsmasq-dns-d6cd869d9-km27m\" (UID: \"30fd93eb-4baf-471e-9dd3-28a623a2a954\") " pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.239328 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.388450 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.544146 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc\") pod \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.544720 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbh7s\" (UniqueName: \"kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s\") pod \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.544797 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb\") pod \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.544870 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb\") pod \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.544955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config\") pod \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\" (UID: \"e6447897-6e23-41ff-a9a4-7c589fdb8d44\") " Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.548248 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s" (OuterVolumeSpecName: "kube-api-access-pbh7s") pod "e6447897-6e23-41ff-a9a4-7c589fdb8d44" (UID: "e6447897-6e23-41ff-a9a4-7c589fdb8d44"). InnerVolumeSpecName "kube-api-access-pbh7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.598701 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e6447897-6e23-41ff-a9a4-7c589fdb8d44" (UID: "e6447897-6e23-41ff-a9a4-7c589fdb8d44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.608326 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config" (OuterVolumeSpecName: "config") pod "e6447897-6e23-41ff-a9a4-7c589fdb8d44" (UID: "e6447897-6e23-41ff-a9a4-7c589fdb8d44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.611720 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e6447897-6e23-41ff-a9a4-7c589fdb8d44" (UID: "e6447897-6e23-41ff-a9a4-7c589fdb8d44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.616607 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e6447897-6e23-41ff-a9a4-7c589fdb8d44" (UID: "e6447897-6e23-41ff-a9a4-7c589fdb8d44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.647598 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.647634 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbh7s\" (UniqueName: \"kubernetes.io/projected/e6447897-6e23-41ff-a9a4-7c589fdb8d44-kube-api-access-pbh7s\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.647651 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.647667 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.647679 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6447897-6e23-41ff-a9a4-7c589fdb8d44-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.752685 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6cd869d9-km27m"] Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.937729 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" event={"ID":"30fd93eb-4baf-471e-9dd3-28a623a2a954","Type":"ContainerStarted","Data":"61ea91a8e51dcef4bd359d27ff5eb5312539e9182cb51950c11379582fc225d7"} Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.950057 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" event={"ID":"e6447897-6e23-41ff-a9a4-7c589fdb8d44","Type":"ContainerDied","Data":"0ad4bd39aae5835771eaf34349083c3511ce0c7d0757230b03e9bfd373ddd01e"} Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.950109 4959 scope.go:117] "RemoveContainer" containerID="473beaff3bab8d1997baebfc4ba75658031b86e24da84ee1ead4015132585ec9" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.950160 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55586cc989-q6ngx" Oct 03 15:20:28 crc kubenswrapper[4959]: I1003 15:20:28.991485 4959 scope.go:117] "RemoveContainer" containerID="94453bf503d05b3330efae4aa0368a92997bf88379d0c2a44c3fc0200860ea45" Oct 03 15:20:29 crc kubenswrapper[4959]: I1003 15:20:29.016388 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:20:29 crc kubenswrapper[4959]: I1003 15:20:29.032827 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55586cc989-q6ngx"] Oct 03 15:20:29 crc kubenswrapper[4959]: I1003 15:20:29.700493 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" path="/var/lib/kubelet/pods/e6447897-6e23-41ff-a9a4-7c589fdb8d44/volumes" Oct 03 15:20:29 crc kubenswrapper[4959]: I1003 15:20:29.963655 4959 generic.go:334] "Generic (PLEG): container finished" podID="30fd93eb-4baf-471e-9dd3-28a623a2a954" containerID="1257c82befeb5e1f5059adbaac329d3ceb711ab917016c2cab77858da6ce903b" exitCode=0 Oct 03 15:20:29 crc kubenswrapper[4959]: I1003 15:20:29.963744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" event={"ID":"30fd93eb-4baf-471e-9dd3-28a623a2a954","Type":"ContainerDied","Data":"1257c82befeb5e1f5059adbaac329d3ceb711ab917016c2cab77858da6ce903b"} Oct 03 15:20:31 crc kubenswrapper[4959]: I1003 15:20:31.009518 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" event={"ID":"30fd93eb-4baf-471e-9dd3-28a623a2a954","Type":"ContainerStarted","Data":"208ad8bbbb5f7bac2ba11685a84534da683acfc3126187cb4db094711e1f15e6"} Oct 03 15:20:31 crc kubenswrapper[4959]: I1003 15:20:31.012534 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:31 crc kubenswrapper[4959]: I1003 15:20:31.036738 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" podStartSLOduration=4.036712239 podStartE2EDuration="4.036712239s" podCreationTimestamp="2025-10-03 15:20:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 15:20:31.027226346 +0000 UTC m=+6600.230569793" watchObservedRunningTime="2025-10-03 15:20:31.036712239 +0000 UTC m=+6600.240055666" Oct 03 15:20:35 crc kubenswrapper[4959]: I1003 15:20:35.686471 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:20:35 crc kubenswrapper[4959]: E1003 15:20:35.687416 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:20:38 crc kubenswrapper[4959]: I1003 15:20:38.241814 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d6cd869d9-km27m" Oct 03 15:20:38 crc kubenswrapper[4959]: I1003 15:20:38.327706 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:38 crc kubenswrapper[4959]: I1003 15:20:38.328171 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="dnsmasq-dns" containerID="cri-o://7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504" gracePeriod=10 Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.003321 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.104350 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.104598 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.104855 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.104946 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.105058 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh242\" (UniqueName: \"kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.105170 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc\") pod \"3a256b35-130b-41bd-bdcc-24fc141c08cd\" (UID: \"3a256b35-130b-41bd-bdcc-24fc141c08cd\") " Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.110347 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242" (OuterVolumeSpecName: "kube-api-access-gh242") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "kube-api-access-gh242". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.111247 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerID="7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504" exitCode=0 Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.111291 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" event={"ID":"3a256b35-130b-41bd-bdcc-24fc141c08cd","Type":"ContainerDied","Data":"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504"} Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.111320 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" event={"ID":"3a256b35-130b-41bd-bdcc-24fc141c08cd","Type":"ContainerDied","Data":"b70bb9dcb1b793929d4236dd805781e2b466dd8c41b5595089c59b858529faed"} Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.111340 4959 scope.go:117] "RemoveContainer" containerID="7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.111498 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b54c866bc-qzbbr" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.167942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.171000 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config" (OuterVolumeSpecName: "config") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.188384 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.189159 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.190900 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3a256b35-130b-41bd-bdcc-24fc141c08cd" (UID: "3a256b35-130b-41bd-bdcc-24fc141c08cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210489 4959 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-config\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210522 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210535 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh242\" (UniqueName: \"kubernetes.io/projected/3a256b35-130b-41bd-bdcc-24fc141c08cd-kube-api-access-gh242\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210545 4959 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210554 4959 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.210562 4959 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a256b35-130b-41bd-bdcc-24fc141c08cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.255900 4959 scope.go:117] "RemoveContainer" containerID="eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.279524 4959 scope.go:117] "RemoveContainer" containerID="7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504" Oct 03 15:20:39 crc kubenswrapper[4959]: E1003 15:20:39.279957 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504\": container with ID starting with 7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504 not found: ID does not exist" containerID="7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.279990 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504"} err="failed to get container status \"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504\": rpc error: code = NotFound desc = could not find container \"7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504\": container with ID starting with 7527f00abe1b5da2fcfd656ea7db1533c49f183bbf02cd39d57c9aa6db935504 not found: ID does not exist" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.280011 4959 scope.go:117] "RemoveContainer" containerID="eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6" Oct 03 15:20:39 crc kubenswrapper[4959]: E1003 15:20:39.280254 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6\": container with ID starting with eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6 not found: ID does not exist" containerID="eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.280272 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6"} err="failed to get container status \"eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6\": rpc error: code = NotFound desc = could not find container \"eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6\": container with ID starting with eca00340445fe2cc2b26d16b661848cc0707b1848d7add17eb974939f36eeed6 not found: ID does not exist" Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.448760 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.456576 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b54c866bc-qzbbr"] Oct 03 15:20:39 crc kubenswrapper[4959]: I1003 15:20:39.702126 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" path="/var/lib/kubelet/pods/3a256b35-130b-41bd-bdcc-24fc141c08cd/volumes" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.989551 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6"] Oct 03 15:20:48 crc kubenswrapper[4959]: E1003 15:20:48.990586 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="init" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990598 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="init" Oct 03 15:20:48 crc kubenswrapper[4959]: E1003 15:20:48.990616 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990622 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: E1003 15:20:48.990636 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990642 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: E1003 15:20:48.990679 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="init" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990685 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="init" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990873 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a256b35-130b-41bd-bdcc-24fc141c08cd" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.990895 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6447897-6e23-41ff-a9a4-7c589fdb8d44" containerName="dnsmasq-dns" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.991651 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.995228 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.995334 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.995530 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:20:48 crc kubenswrapper[4959]: I1003 15:20:48.995548 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.010406 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6"] Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.059295 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.059362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.059512 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.059677 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.059719 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgxzd\" (UniqueName: \"kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.162301 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.162360 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.162409 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.162470 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.162498 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgxzd\" (UniqueName: \"kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.168175 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.168262 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.168403 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.168486 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.192136 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgxzd\" (UniqueName: \"kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.366138 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:20:49 crc kubenswrapper[4959]: I1003 15:20:49.685971 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:20:49 crc kubenswrapper[4959]: E1003 15:20:49.686453 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:20:50 crc kubenswrapper[4959]: W1003 15:20:50.100748 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21dadec9_3ba5_4010_a9ea_e65cbe8cbca3.slice/crio-99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89 WatchSource:0}: Error finding container 99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89: Status 404 returned error can't find the container with id 99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89 Oct 03 15:20:50 crc kubenswrapper[4959]: I1003 15:20:50.104066 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6"] Oct 03 15:20:50 crc kubenswrapper[4959]: I1003 15:20:50.104286 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:20:50 crc kubenswrapper[4959]: I1003 15:20:50.242502 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" event={"ID":"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3","Type":"ContainerStarted","Data":"99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89"} Oct 03 15:21:01 crc kubenswrapper[4959]: I1003 15:21:01.692817 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:21:01 crc kubenswrapper[4959]: E1003 15:21:01.693799 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:21:05 crc kubenswrapper[4959]: I1003 15:21:05.416900 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" event={"ID":"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3","Type":"ContainerStarted","Data":"b55ad8ac273d3b3df3156286264393fd58b4c04bdfc81d39cff0de40e0fd3795"} Oct 03 15:21:05 crc kubenswrapper[4959]: I1003 15:21:05.441320 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" podStartSLOduration=3.128530176 podStartE2EDuration="17.441289338s" podCreationTimestamp="2025-10-03 15:20:48 +0000 UTC" firstStartedPulling="2025-10-03 15:20:50.103984954 +0000 UTC m=+6619.307328371" lastFinishedPulling="2025-10-03 15:21:04.416744106 +0000 UTC m=+6633.620087533" observedRunningTime="2025-10-03 15:21:05.437611959 +0000 UTC m=+6634.640955376" watchObservedRunningTime="2025-10-03 15:21:05.441289338 +0000 UTC m=+6634.644632755" Oct 03 15:21:10 crc kubenswrapper[4959]: I1003 15:21:10.028038 4959 scope.go:117] "RemoveContainer" containerID="6b12290d281b9f0de0443839cfc0ff13ded37637f2c656c14146753fe716bbab" Oct 03 15:21:10 crc kubenswrapper[4959]: I1003 15:21:10.246878 4959 scope.go:117] "RemoveContainer" containerID="9d87c1ccae0832808d108bad5ef0b40207f9632d771db885a7a866bb6cebec12" Oct 03 15:21:12 crc kubenswrapper[4959]: I1003 15:21:12.685682 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:21:12 crc kubenswrapper[4959]: E1003 15:21:12.686492 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:21:18 crc kubenswrapper[4959]: I1003 15:21:18.577594 4959 generic.go:334] "Generic (PLEG): container finished" podID="21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" containerID="b55ad8ac273d3b3df3156286264393fd58b4c04bdfc81d39cff0de40e0fd3795" exitCode=0 Oct 03 15:21:18 crc kubenswrapper[4959]: I1003 15:21:18.577681 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" event={"ID":"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3","Type":"ContainerDied","Data":"b55ad8ac273d3b3df3156286264393fd58b4c04bdfc81d39cff0de40e0fd3795"} Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.167146 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.282417 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key\") pod \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.282502 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph\") pod \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.282721 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory\") pod \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.282936 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgxzd\" (UniqueName: \"kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd\") pod \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.283012 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle\") pod \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\" (UID: \"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3\") " Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.290458 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd" (OuterVolumeSpecName: "kube-api-access-zgxzd") pod "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" (UID: "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3"). InnerVolumeSpecName "kube-api-access-zgxzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.291307 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" (UID: "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.291474 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph" (OuterVolumeSpecName: "ceph") pod "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" (UID: "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.330397 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory" (OuterVolumeSpecName: "inventory") pod "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" (UID: "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.330948 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" (UID: "21dadec9-3ba5-4010-a9ea-e65cbe8cbca3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.385460 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.385504 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgxzd\" (UniqueName: \"kubernetes.io/projected/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-kube-api-access-zgxzd\") on node \"crc\" DevicePath \"\"" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.385516 4959 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.385524 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.385534 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/21dadec9-3ba5-4010-a9ea-e65cbe8cbca3-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.605440 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" event={"ID":"21dadec9-3ba5-4010-a9ea-e65cbe8cbca3","Type":"ContainerDied","Data":"99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89"} Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.605483 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99ab9e66b79cb99ce3412f4dfb3ce63ee5f047629b812bce28cc66394e153f89" Oct 03 15:21:20 crc kubenswrapper[4959]: I1003 15:21:20.605544 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.661573 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g"] Oct 03 15:21:21 crc kubenswrapper[4959]: E1003 15:21:21.662166 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.662183 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.662488 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="21dadec9-3ba5-4010-a9ea-e65cbe8cbca3" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.663478 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.667721 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.667861 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.667989 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.669724 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g"] Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.670097 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.724052 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.724552 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.724760 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.724961 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl5nm\" (UniqueName: \"kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.725127 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.827414 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl5nm\" (UniqueName: \"kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.827466 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.827580 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.827647 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.827682 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.834451 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.834894 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.835375 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.835829 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.847140 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl5nm\" (UniqueName: \"kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:21 crc kubenswrapper[4959]: I1003 15:21:21.990364 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:21:22 crc kubenswrapper[4959]: I1003 15:21:22.669090 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g"] Oct 03 15:21:23 crc kubenswrapper[4959]: I1003 15:21:23.654155 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" event={"ID":"034c6ab7-82c8-4922-ae56-e1862f88d39e","Type":"ContainerStarted","Data":"2b64463677badd0f46938591beb1d06e335cb4e1e9d2426422732b330a23fd46"} Oct 03 15:21:23 crc kubenswrapper[4959]: I1003 15:21:23.655080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" event={"ID":"034c6ab7-82c8-4922-ae56-e1862f88d39e","Type":"ContainerStarted","Data":"b2c2da22d2f828cd6df5eaf305d4e45dea7551fdb7feb2d98cabe47f7b1ff390"} Oct 03 15:21:23 crc kubenswrapper[4959]: I1003 15:21:23.682517 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" podStartSLOduration=2.335831334 podStartE2EDuration="2.682498667s" podCreationTimestamp="2025-10-03 15:21:21 +0000 UTC" firstStartedPulling="2025-10-03 15:21:22.682997197 +0000 UTC m=+6651.886340614" lastFinishedPulling="2025-10-03 15:21:23.02966453 +0000 UTC m=+6652.233007947" observedRunningTime="2025-10-03 15:21:23.678346865 +0000 UTC m=+6652.881690322" watchObservedRunningTime="2025-10-03 15:21:23.682498667 +0000 UTC m=+6652.885842084" Oct 03 15:21:25 crc kubenswrapper[4959]: I1003 15:21:25.689516 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:21:25 crc kubenswrapper[4959]: E1003 15:21:25.690152 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:21:37 crc kubenswrapper[4959]: I1003 15:21:37.687417 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:21:38 crc kubenswrapper[4959]: I1003 15:21:38.831186 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b"} Oct 03 15:21:47 crc kubenswrapper[4959]: I1003 15:21:47.054490 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-js7ld"] Oct 03 15:21:47 crc kubenswrapper[4959]: I1003 15:21:47.066829 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-js7ld"] Oct 03 15:21:47 crc kubenswrapper[4959]: I1003 15:21:47.702871 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c285ce72-79d4-44b8-aa1a-f45d6b595bab" path="/var/lib/kubelet/pods/c285ce72-79d4-44b8-aa1a-f45d6b595bab/volumes" Oct 03 15:21:59 crc kubenswrapper[4959]: I1003 15:21:59.045668 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-e9e2-account-create-dkfjk"] Oct 03 15:21:59 crc kubenswrapper[4959]: I1003 15:21:59.065934 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-e9e2-account-create-dkfjk"] Oct 03 15:21:59 crc kubenswrapper[4959]: I1003 15:21:59.696113 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de58b2a6-3503-4b49-bf4a-c393087c4f47" path="/var/lib/kubelet/pods/de58b2a6-3503-4b49-bf4a-c393087c4f47/volumes" Oct 03 15:22:05 crc kubenswrapper[4959]: I1003 15:22:05.028654 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-tvbsr"] Oct 03 15:22:05 crc kubenswrapper[4959]: I1003 15:22:05.039307 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-tvbsr"] Oct 03 15:22:05 crc kubenswrapper[4959]: I1003 15:22:05.699289 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="670af6f4-13ae-406a-b0db-8a84ada6ea30" path="/var/lib/kubelet/pods/670af6f4-13ae-406a-b0db-8a84ada6ea30/volumes" Oct 03 15:22:10 crc kubenswrapper[4959]: I1003 15:22:10.364044 4959 scope.go:117] "RemoveContainer" containerID="eaaa035a20b4fa311f28e051b156b075223070ad9a7064fd55139278380ee8ac" Oct 03 15:22:10 crc kubenswrapper[4959]: I1003 15:22:10.407232 4959 scope.go:117] "RemoveContainer" containerID="91773be05ca8ae836e57c8108eebfd3b9616b4a689b4db37cdbe8fb9e63a9fea" Oct 03 15:22:10 crc kubenswrapper[4959]: I1003 15:22:10.529590 4959 scope.go:117] "RemoveContainer" containerID="4b0786610bbe66998aa9ab2a84ef623f2abf04f98f9ef3aa67a70e6363d09933" Oct 03 15:22:16 crc kubenswrapper[4959]: I1003 15:22:16.046852 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-d0ae-account-create-cf4c4"] Oct 03 15:22:16 crc kubenswrapper[4959]: I1003 15:22:16.057239 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-d0ae-account-create-cf4c4"] Oct 03 15:22:17 crc kubenswrapper[4959]: I1003 15:22:17.709854 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c90709-b50f-4c10-af62-f646f51fad07" path="/var/lib/kubelet/pods/f8c90709-b50f-4c10-af62-f646f51fad07/volumes" Oct 03 15:22:51 crc kubenswrapper[4959]: I1003 15:22:51.076362 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-n5bzq"] Oct 03 15:22:51 crc kubenswrapper[4959]: I1003 15:22:51.092212 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-n5bzq"] Oct 03 15:22:51 crc kubenswrapper[4959]: I1003 15:22:51.715665 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f977e88e-bc91-4efc-ac44-686076979802" path="/var/lib/kubelet/pods/f977e88e-bc91-4efc-ac44-686076979802/volumes" Oct 03 15:23:10 crc kubenswrapper[4959]: I1003 15:23:10.637060 4959 scope.go:117] "RemoveContainer" containerID="a1e57e3b0e723aa9b56db1a4491d4a3cc255c2c37aa2b0192b5d764d3ac1f475" Oct 03 15:23:10 crc kubenswrapper[4959]: I1003 15:23:10.671701 4959 scope.go:117] "RemoveContainer" containerID="f36e0adb5e6ed0a3a1b49a17c906145ac6bfe2a285849bb8a6fea137377e17ef" Oct 03 15:23:10 crc kubenswrapper[4959]: I1003 15:23:10.750273 4959 scope.go:117] "RemoveContainer" containerID="316abf3d476478c428b230de8b4a497af167965d7b3f3a6fa83402d95e383d1f" Oct 03 15:23:38 crc kubenswrapper[4959]: I1003 15:23:38.873458 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:38 crc kubenswrapper[4959]: I1003 15:23:38.877922 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:38 crc kubenswrapper[4959]: I1003 15:23:38.884058 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.035251 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.035319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfj9\" (UniqueName: \"kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.035395 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.137204 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.137263 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfj9\" (UniqueName: \"kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.137337 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.137920 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.138138 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.156605 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfj9\" (UniqueName: \"kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9\") pod \"community-operators-w4hp6\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.199958 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:39 crc kubenswrapper[4959]: I1003 15:23:39.739432 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:40 crc kubenswrapper[4959]: I1003 15:23:40.247658 4959 generic.go:334] "Generic (PLEG): container finished" podID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerID="c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214" exitCode=0 Oct 03 15:23:40 crc kubenswrapper[4959]: I1003 15:23:40.247725 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerDied","Data":"c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214"} Oct 03 15:23:40 crc kubenswrapper[4959]: I1003 15:23:40.247962 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerStarted","Data":"48bed159f9588d7e726e0d0d7ab8dca300878af2fea258f39d443b243d05804e"} Oct 03 15:23:42 crc kubenswrapper[4959]: I1003 15:23:42.272516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerStarted","Data":"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236"} Oct 03 15:23:43 crc kubenswrapper[4959]: I1003 15:23:43.285095 4959 generic.go:334] "Generic (PLEG): container finished" podID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerID="162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236" exitCode=0 Oct 03 15:23:43 crc kubenswrapper[4959]: I1003 15:23:43.285230 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerDied","Data":"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236"} Oct 03 15:23:44 crc kubenswrapper[4959]: I1003 15:23:44.297516 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerStarted","Data":"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe"} Oct 03 15:23:44 crc kubenswrapper[4959]: I1003 15:23:44.321148 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w4hp6" podStartSLOduration=2.824939601 podStartE2EDuration="6.32112748s" podCreationTimestamp="2025-10-03 15:23:38 +0000 UTC" firstStartedPulling="2025-10-03 15:23:40.250452261 +0000 UTC m=+6789.453795678" lastFinishedPulling="2025-10-03 15:23:43.74664013 +0000 UTC m=+6792.949983557" observedRunningTime="2025-10-03 15:23:44.318121876 +0000 UTC m=+6793.521465313" watchObservedRunningTime="2025-10-03 15:23:44.32112748 +0000 UTC m=+6793.524470897" Oct 03 15:23:49 crc kubenswrapper[4959]: I1003 15:23:49.200530 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:49 crc kubenswrapper[4959]: I1003 15:23:49.201459 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:49 crc kubenswrapper[4959]: I1003 15:23:49.291622 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:49 crc kubenswrapper[4959]: I1003 15:23:49.447809 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:49 crc kubenswrapper[4959]: I1003 15:23:49.564768 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:51 crc kubenswrapper[4959]: I1003 15:23:51.412138 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w4hp6" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="registry-server" containerID="cri-o://317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe" gracePeriod=2 Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.074314 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.244078 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdfj9\" (UniqueName: \"kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9\") pod \"5bc849ad-d9c7-47da-981c-d32c3122c447\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.244153 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities\") pod \"5bc849ad-d9c7-47da-981c-d32c3122c447\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.244481 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content\") pod \"5bc849ad-d9c7-47da-981c-d32c3122c447\" (UID: \"5bc849ad-d9c7-47da-981c-d32c3122c447\") " Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.245000 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities" (OuterVolumeSpecName: "utilities") pod "5bc849ad-d9c7-47da-981c-d32c3122c447" (UID: "5bc849ad-d9c7-47da-981c-d32c3122c447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.246471 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.255187 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9" (OuterVolumeSpecName: "kube-api-access-sdfj9") pod "5bc849ad-d9c7-47da-981c-d32c3122c447" (UID: "5bc849ad-d9c7-47da-981c-d32c3122c447"). InnerVolumeSpecName "kube-api-access-sdfj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.290880 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bc849ad-d9c7-47da-981c-d32c3122c447" (UID: "5bc849ad-d9c7-47da-981c-d32c3122c447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.349474 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc849ad-d9c7-47da-981c-d32c3122c447-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.349527 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdfj9\" (UniqueName: \"kubernetes.io/projected/5bc849ad-d9c7-47da-981c-d32c3122c447-kube-api-access-sdfj9\") on node \"crc\" DevicePath \"\"" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.424124 4959 generic.go:334] "Generic (PLEG): container finished" podID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerID="317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe" exitCode=0 Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.424162 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w4hp6" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.424182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerDied","Data":"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe"} Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.424421 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w4hp6" event={"ID":"5bc849ad-d9c7-47da-981c-d32c3122c447","Type":"ContainerDied","Data":"48bed159f9588d7e726e0d0d7ab8dca300878af2fea258f39d443b243d05804e"} Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.424462 4959 scope.go:117] "RemoveContainer" containerID="317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.456449 4959 scope.go:117] "RemoveContainer" containerID="162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.493167 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.498507 4959 scope.go:117] "RemoveContainer" containerID="c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.498789 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w4hp6"] Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.552892 4959 scope.go:117] "RemoveContainer" containerID="317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe" Oct 03 15:23:52 crc kubenswrapper[4959]: E1003 15:23:52.553499 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe\": container with ID starting with 317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe not found: ID does not exist" containerID="317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.553525 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe"} err="failed to get container status \"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe\": rpc error: code = NotFound desc = could not find container \"317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe\": container with ID starting with 317f7f8ad135ac79b5844b635a5389850a4b21f7dc6c797036aa65fbe2c88ebe not found: ID does not exist" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.553731 4959 scope.go:117] "RemoveContainer" containerID="162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236" Oct 03 15:23:52 crc kubenswrapper[4959]: E1003 15:23:52.554382 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236\": container with ID starting with 162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236 not found: ID does not exist" containerID="162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.554407 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236"} err="failed to get container status \"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236\": rpc error: code = NotFound desc = could not find container \"162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236\": container with ID starting with 162cfc52df9262d34af7fd3a7a1725b6ce5606a7c318ede1d99ebe48bab1f236 not found: ID does not exist" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.554423 4959 scope.go:117] "RemoveContainer" containerID="c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214" Oct 03 15:23:52 crc kubenswrapper[4959]: E1003 15:23:52.554820 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214\": container with ID starting with c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214 not found: ID does not exist" containerID="c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214" Oct 03 15:23:52 crc kubenswrapper[4959]: I1003 15:23:52.554868 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214"} err="failed to get container status \"c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214\": rpc error: code = NotFound desc = could not find container \"c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214\": container with ID starting with c0e92bfa4e6edc2c2c789947ab3b5635d8c120c4d14b82a6c2af292340f68214 not found: ID does not exist" Oct 03 15:23:53 crc kubenswrapper[4959]: I1003 15:23:53.696513 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" path="/var/lib/kubelet/pods/5bc849ad-d9c7-47da-981c-d32c3122c447/volumes" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.691661 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:04 crc kubenswrapper[4959]: E1003 15:24:04.692685 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="registry-server" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.692701 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="registry-server" Oct 03 15:24:04 crc kubenswrapper[4959]: E1003 15:24:04.692718 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="extract-content" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.692729 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="extract-content" Oct 03 15:24:04 crc kubenswrapper[4959]: E1003 15:24:04.692756 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="extract-utilities" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.692764 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="extract-utilities" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.693031 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc849ad-d9c7-47da-981c-d32c3122c447" containerName="registry-server" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.694977 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.705862 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.826592 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.826936 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbkzg\" (UniqueName: \"kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.826987 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.929326 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.929760 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.929966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbkzg\" (UniqueName: \"kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.929998 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.930309 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:04 crc kubenswrapper[4959]: I1003 15:24:04.954065 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbkzg\" (UniqueName: \"kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg\") pod \"redhat-operators-frdzt\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:05 crc kubenswrapper[4959]: I1003 15:24:05.018738 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:05 crc kubenswrapper[4959]: I1003 15:24:05.489872 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:05 crc kubenswrapper[4959]: I1003 15:24:05.574113 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerStarted","Data":"89fd64cdc60764e3b35edbac96e63afed420c5ae70f130bd10356d1a32acae05"} Oct 03 15:24:06 crc kubenswrapper[4959]: I1003 15:24:06.044058 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:24:06 crc kubenswrapper[4959]: I1003 15:24:06.044393 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:24:06 crc kubenswrapper[4959]: I1003 15:24:06.587479 4959 generic.go:334] "Generic (PLEG): container finished" podID="3caa0951-a623-4cce-903a-6953c22c5282" containerID="126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41" exitCode=0 Oct 03 15:24:06 crc kubenswrapper[4959]: I1003 15:24:06.587522 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerDied","Data":"126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41"} Oct 03 15:24:08 crc kubenswrapper[4959]: I1003 15:24:08.605491 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerStarted","Data":"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06"} Oct 03 15:24:12 crc kubenswrapper[4959]: E1003 15:24:12.106819 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3caa0951_a623_4cce_903a_6953c22c5282.slice/crio-2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06.scope\": RecentStats: unable to find data in memory cache]" Oct 03 15:24:12 crc kubenswrapper[4959]: I1003 15:24:12.654997 4959 generic.go:334] "Generic (PLEG): container finished" podID="3caa0951-a623-4cce-903a-6953c22c5282" containerID="2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06" exitCode=0 Oct 03 15:24:12 crc kubenswrapper[4959]: I1003 15:24:12.655075 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerDied","Data":"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06"} Oct 03 15:24:13 crc kubenswrapper[4959]: I1003 15:24:13.674505 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerStarted","Data":"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb"} Oct 03 15:24:13 crc kubenswrapper[4959]: I1003 15:24:13.700930 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frdzt" podStartSLOduration=3.120447854 podStartE2EDuration="9.700908951s" podCreationTimestamp="2025-10-03 15:24:04 +0000 UTC" firstStartedPulling="2025-10-03 15:24:06.596043471 +0000 UTC m=+6815.799386918" lastFinishedPulling="2025-10-03 15:24:13.176504588 +0000 UTC m=+6822.379848015" observedRunningTime="2025-10-03 15:24:13.697708423 +0000 UTC m=+6822.901051860" watchObservedRunningTime="2025-10-03 15:24:13.700908951 +0000 UTC m=+6822.904252368" Oct 03 15:24:15 crc kubenswrapper[4959]: I1003 15:24:15.018922 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:15 crc kubenswrapper[4959]: I1003 15:24:15.019435 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:16 crc kubenswrapper[4959]: I1003 15:24:16.092581 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frdzt" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" probeResult="failure" output=< Oct 03 15:24:16 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:24:16 crc kubenswrapper[4959]: > Oct 03 15:24:26 crc kubenswrapper[4959]: I1003 15:24:26.086829 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frdzt" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" probeResult="failure" output=< Oct 03 15:24:26 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:24:26 crc kubenswrapper[4959]: > Oct 03 15:24:36 crc kubenswrapper[4959]: I1003 15:24:36.044360 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:24:36 crc kubenswrapper[4959]: I1003 15:24:36.044999 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:24:36 crc kubenswrapper[4959]: I1003 15:24:36.100895 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frdzt" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" probeResult="failure" output=< Oct 03 15:24:36 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:24:36 crc kubenswrapper[4959]: > Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.525394 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.528182 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.535423 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.674976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.675021 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.675964 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chs4s\" (UniqueName: \"kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.778596 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chs4s\" (UniqueName: \"kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.778739 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.778763 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.779314 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.779313 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.813385 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chs4s\" (UniqueName: \"kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s\") pod \"redhat-marketplace-fpm8s\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:37 crc kubenswrapper[4959]: I1003 15:24:37.862133 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:38 crc kubenswrapper[4959]: I1003 15:24:38.337392 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:24:38 crc kubenswrapper[4959]: I1003 15:24:38.972354 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerID="fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0" exitCode=0 Oct 03 15:24:38 crc kubenswrapper[4959]: I1003 15:24:38.972424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerDied","Data":"fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0"} Oct 03 15:24:38 crc kubenswrapper[4959]: I1003 15:24:38.972684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerStarted","Data":"07d9e5ea016a49da4287408075289fb633851c3eca6ea3f0c48da835b6fb309e"} Oct 03 15:24:39 crc kubenswrapper[4959]: I1003 15:24:39.984050 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerStarted","Data":"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28"} Oct 03 15:24:43 crc kubenswrapper[4959]: I1003 15:24:43.016129 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerID="4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28" exitCode=0 Oct 03 15:24:43 crc kubenswrapper[4959]: I1003 15:24:43.016201 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerDied","Data":"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28"} Oct 03 15:24:45 crc kubenswrapper[4959]: I1003 15:24:45.038468 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerStarted","Data":"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c"} Oct 03 15:24:45 crc kubenswrapper[4959]: I1003 15:24:45.070446 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fpm8s" podStartSLOduration=3.157708784 podStartE2EDuration="8.070420306s" podCreationTimestamp="2025-10-03 15:24:37 +0000 UTC" firstStartedPulling="2025-10-03 15:24:38.977158562 +0000 UTC m=+6848.180501989" lastFinishedPulling="2025-10-03 15:24:43.889870094 +0000 UTC m=+6853.093213511" observedRunningTime="2025-10-03 15:24:45.059607811 +0000 UTC m=+6854.262951258" watchObservedRunningTime="2025-10-03 15:24:45.070420306 +0000 UTC m=+6854.273763753" Oct 03 15:24:45 crc kubenswrapper[4959]: I1003 15:24:45.075275 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:45 crc kubenswrapper[4959]: I1003 15:24:45.136764 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:46 crc kubenswrapper[4959]: I1003 15:24:46.274965 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.069344 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frdzt" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" containerID="cri-o://00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb" gracePeriod=2 Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.662748 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.836723 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content\") pod \"3caa0951-a623-4cce-903a-6953c22c5282\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.837241 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbkzg\" (UniqueName: \"kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg\") pod \"3caa0951-a623-4cce-903a-6953c22c5282\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.837285 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities\") pod \"3caa0951-a623-4cce-903a-6953c22c5282\" (UID: \"3caa0951-a623-4cce-903a-6953c22c5282\") " Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.838262 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities" (OuterVolumeSpecName: "utilities") pod "3caa0951-a623-4cce-903a-6953c22c5282" (UID: "3caa0951-a623-4cce-903a-6953c22c5282"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.842473 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg" (OuterVolumeSpecName: "kube-api-access-dbkzg") pod "3caa0951-a623-4cce-903a-6953c22c5282" (UID: "3caa0951-a623-4cce-903a-6953c22c5282"). InnerVolumeSpecName "kube-api-access-dbkzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.863296 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.863348 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.912186 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3caa0951-a623-4cce-903a-6953c22c5282" (UID: "3caa0951-a623-4cce-903a-6953c22c5282"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.939972 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbkzg\" (UniqueName: \"kubernetes.io/projected/3caa0951-a623-4cce-903a-6953c22c5282-kube-api-access-dbkzg\") on node \"crc\" DevicePath \"\"" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.940007 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:24:47 crc kubenswrapper[4959]: I1003 15:24:47.940020 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa0951-a623-4cce-903a-6953c22c5282-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.080023 4959 generic.go:334] "Generic (PLEG): container finished" podID="3caa0951-a623-4cce-903a-6953c22c5282" containerID="00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb" exitCode=0 Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.080074 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerDied","Data":"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb"} Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.080105 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frdzt" event={"ID":"3caa0951-a623-4cce-903a-6953c22c5282","Type":"ContainerDied","Data":"89fd64cdc60764e3b35edbac96e63afed420c5ae70f130bd10356d1a32acae05"} Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.080128 4959 scope.go:117] "RemoveContainer" containerID="00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.080300 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frdzt" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.120114 4959 scope.go:117] "RemoveContainer" containerID="2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.120239 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.127707 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frdzt"] Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.145261 4959 scope.go:117] "RemoveContainer" containerID="126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.199786 4959 scope.go:117] "RemoveContainer" containerID="00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb" Oct 03 15:24:48 crc kubenswrapper[4959]: E1003 15:24:48.200161 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb\": container with ID starting with 00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb not found: ID does not exist" containerID="00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.200224 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb"} err="failed to get container status \"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb\": rpc error: code = NotFound desc = could not find container \"00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb\": container with ID starting with 00b95355b760903da70a3d0067920bc6d8eea342b561f6681912d8281a88a8eb not found: ID does not exist" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.200258 4959 scope.go:117] "RemoveContainer" containerID="2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06" Oct 03 15:24:48 crc kubenswrapper[4959]: E1003 15:24:48.200669 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06\": container with ID starting with 2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06 not found: ID does not exist" containerID="2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.200705 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06"} err="failed to get container status \"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06\": rpc error: code = NotFound desc = could not find container \"2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06\": container with ID starting with 2de01c3933c393bc94b17e21285fbb1dd29ffa3b21f9634efd50a7264a9c5e06 not found: ID does not exist" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.200732 4959 scope.go:117] "RemoveContainer" containerID="126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41" Oct 03 15:24:48 crc kubenswrapper[4959]: E1003 15:24:48.200981 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41\": container with ID starting with 126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41 not found: ID does not exist" containerID="126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.201015 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41"} err="failed to get container status \"126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41\": rpc error: code = NotFound desc = could not find container \"126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41\": container with ID starting with 126aaf7d649f7172c9acc3e7987468efd6e7de2a5e9b054e920dc3a818baac41 not found: ID does not exist" Oct 03 15:24:48 crc kubenswrapper[4959]: I1003 15:24:48.909677 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-fpm8s" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="registry-server" probeResult="failure" output=< Oct 03 15:24:48 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:24:48 crc kubenswrapper[4959]: > Oct 03 15:24:49 crc kubenswrapper[4959]: I1003 15:24:49.705279 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3caa0951-a623-4cce-903a-6953c22c5282" path="/var/lib/kubelet/pods/3caa0951-a623-4cce-903a-6953c22c5282/volumes" Oct 03 15:24:57 crc kubenswrapper[4959]: I1003 15:24:57.913311 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:57 crc kubenswrapper[4959]: I1003 15:24:57.964767 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:58 crc kubenswrapper[4959]: I1003 15:24:58.153093 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.191851 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fpm8s" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="registry-server" containerID="cri-o://7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c" gracePeriod=2 Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.741421 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.909582 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chs4s\" (UniqueName: \"kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s\") pod \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.909836 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content\") pod \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.909917 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities\") pod \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\" (UID: \"aa2540ac-1a89-42ff-ac1e-1897de3d6f11\") " Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.911341 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities" (OuterVolumeSpecName: "utilities") pod "aa2540ac-1a89-42ff-ac1e-1897de3d6f11" (UID: "aa2540ac-1a89-42ff-ac1e-1897de3d6f11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.915062 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s" (OuterVolumeSpecName: "kube-api-access-chs4s") pod "aa2540ac-1a89-42ff-ac1e-1897de3d6f11" (UID: "aa2540ac-1a89-42ff-ac1e-1897de3d6f11"). InnerVolumeSpecName "kube-api-access-chs4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:24:59 crc kubenswrapper[4959]: I1003 15:24:59.923103 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa2540ac-1a89-42ff-ac1e-1897de3d6f11" (UID: "aa2540ac-1a89-42ff-ac1e-1897de3d6f11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.012213 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chs4s\" (UniqueName: \"kubernetes.io/projected/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-kube-api-access-chs4s\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.012288 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.012303 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa2540ac-1a89-42ff-ac1e-1897de3d6f11-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.207453 4959 generic.go:334] "Generic (PLEG): container finished" podID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerID="7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c" exitCode=0 Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.207538 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerDied","Data":"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c"} Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.207607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fpm8s" event={"ID":"aa2540ac-1a89-42ff-ac1e-1897de3d6f11","Type":"ContainerDied","Data":"07d9e5ea016a49da4287408075289fb633851c3eca6ea3f0c48da835b6fb309e"} Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.207549 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fpm8s" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.207644 4959 scope.go:117] "RemoveContainer" containerID="7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.247271 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.249802 4959 scope.go:117] "RemoveContainer" containerID="4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.255822 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fpm8s"] Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.273239 4959 scope.go:117] "RemoveContainer" containerID="fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.333333 4959 scope.go:117] "RemoveContainer" containerID="7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c" Oct 03 15:25:00 crc kubenswrapper[4959]: E1003 15:25:00.333829 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c\": container with ID starting with 7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c not found: ID does not exist" containerID="7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.333895 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c"} err="failed to get container status \"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c\": rpc error: code = NotFound desc = could not find container \"7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c\": container with ID starting with 7c385172791858dac92b0458d4beaf15ecf60b64ebd59d76c9acb08b496f250c not found: ID does not exist" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.333924 4959 scope.go:117] "RemoveContainer" containerID="4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28" Oct 03 15:25:00 crc kubenswrapper[4959]: E1003 15:25:00.334453 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28\": container with ID starting with 4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28 not found: ID does not exist" containerID="4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.334481 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28"} err="failed to get container status \"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28\": rpc error: code = NotFound desc = could not find container \"4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28\": container with ID starting with 4d7c79e02b029be4cfe209ac25da2057832bd22c51ac18c54cdbe46419a6fe28 not found: ID does not exist" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.334499 4959 scope.go:117] "RemoveContainer" containerID="fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0" Oct 03 15:25:00 crc kubenswrapper[4959]: E1003 15:25:00.334770 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0\": container with ID starting with fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0 not found: ID does not exist" containerID="fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0" Oct 03 15:25:00 crc kubenswrapper[4959]: I1003 15:25:00.334795 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0"} err="failed to get container status \"fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0\": rpc error: code = NotFound desc = could not find container \"fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0\": container with ID starting with fccad02d7503d6f459934c3ca6d5375a149a7b285ef8412d141cccaed2a7bdb0 not found: ID does not exist" Oct 03 15:25:01 crc kubenswrapper[4959]: I1003 15:25:01.711002 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" path="/var/lib/kubelet/pods/aa2540ac-1a89-42ff-ac1e-1897de3d6f11/volumes" Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.044899 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.046025 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.046150 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.048343 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.048524 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b" gracePeriod=600 Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.273659 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b" exitCode=0 Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.273730 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b"} Oct 03 15:25:06 crc kubenswrapper[4959]: I1003 15:25:06.273993 4959 scope.go:117] "RemoveContainer" containerID="39f44efe6d84bd1bdda0854ea32e5b06d2ae411c39fcf16b7e3f5b30cd37f176" Oct 03 15:25:07 crc kubenswrapper[4959]: I1003 15:25:07.283973 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431"} Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.102511 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104421 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="extract-utilities" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104440 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="extract-utilities" Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104484 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104493 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104517 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="extract-content" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104525 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="extract-content" Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104545 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104553 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104566 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="extract-content" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104573 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="extract-content" Oct 03 15:25:24 crc kubenswrapper[4959]: E1003 15:25:24.104591 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="extract-utilities" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104598 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="extract-utilities" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104834 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3caa0951-a623-4cce-903a-6953c22c5282" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.104859 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa2540ac-1a89-42ff-ac1e-1897de3d6f11" containerName="registry-server" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.107992 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.114298 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.263749 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.264081 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.264319 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwlrw\" (UniqueName: \"kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.366821 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.367255 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.367392 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwlrw\" (UniqueName: \"kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.367391 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.367684 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.391288 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwlrw\" (UniqueName: \"kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw\") pod \"certified-operators-jn97x\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.442230 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:24 crc kubenswrapper[4959]: I1003 15:25:24.965004 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:24 crc kubenswrapper[4959]: W1003 15:25:24.980493 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48125534_76d4_4d10_982c_bbc34761886f.slice/crio-bc9b2f3900ecaf20b3b9dd49cffbad40acd1704eed5f71e1b4924e6bfbd7c223 WatchSource:0}: Error finding container bc9b2f3900ecaf20b3b9dd49cffbad40acd1704eed5f71e1b4924e6bfbd7c223: Status 404 returned error can't find the container with id bc9b2f3900ecaf20b3b9dd49cffbad40acd1704eed5f71e1b4924e6bfbd7c223 Oct 03 15:25:25 crc kubenswrapper[4959]: I1003 15:25:25.499821 4959 generic.go:334] "Generic (PLEG): container finished" podID="48125534-76d4-4d10-982c-bbc34761886f" containerID="e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed" exitCode=0 Oct 03 15:25:25 crc kubenswrapper[4959]: I1003 15:25:25.499946 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerDied","Data":"e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed"} Oct 03 15:25:25 crc kubenswrapper[4959]: I1003 15:25:25.500261 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerStarted","Data":"bc9b2f3900ecaf20b3b9dd49cffbad40acd1704eed5f71e1b4924e6bfbd7c223"} Oct 03 15:25:26 crc kubenswrapper[4959]: I1003 15:25:26.516885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerStarted","Data":"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992"} Oct 03 15:25:27 crc kubenswrapper[4959]: I1003 15:25:27.531892 4959 generic.go:334] "Generic (PLEG): container finished" podID="48125534-76d4-4d10-982c-bbc34761886f" containerID="af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992" exitCode=0 Oct 03 15:25:27 crc kubenswrapper[4959]: I1003 15:25:27.532033 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerDied","Data":"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992"} Oct 03 15:25:28 crc kubenswrapper[4959]: I1003 15:25:28.545271 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerStarted","Data":"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b"} Oct 03 15:25:28 crc kubenswrapper[4959]: I1003 15:25:28.575431 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jn97x" podStartSLOduration=2.108564842 podStartE2EDuration="4.575412171s" podCreationTimestamp="2025-10-03 15:25:24 +0000 UTC" firstStartedPulling="2025-10-03 15:25:25.506637651 +0000 UTC m=+6894.709981068" lastFinishedPulling="2025-10-03 15:25:27.97348498 +0000 UTC m=+6897.176828397" observedRunningTime="2025-10-03 15:25:28.569750101 +0000 UTC m=+6897.773093528" watchObservedRunningTime="2025-10-03 15:25:28.575412171 +0000 UTC m=+6897.778755588" Oct 03 15:25:30 crc kubenswrapper[4959]: I1003 15:25:30.039667 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-sbh94"] Oct 03 15:25:30 crc kubenswrapper[4959]: I1003 15:25:30.051039 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-sbh94"] Oct 03 15:25:31 crc kubenswrapper[4959]: I1003 15:25:31.708572 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689" path="/var/lib/kubelet/pods/e4e5b714-c3aa-4c9f-a86a-a65ec0b0d689/volumes" Oct 03 15:25:34 crc kubenswrapper[4959]: I1003 15:25:34.442675 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:34 crc kubenswrapper[4959]: I1003 15:25:34.443092 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:34 crc kubenswrapper[4959]: I1003 15:25:34.521815 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:34 crc kubenswrapper[4959]: I1003 15:25:34.648250 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:34 crc kubenswrapper[4959]: I1003 15:25:34.772225 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:36 crc kubenswrapper[4959]: I1003 15:25:36.630834 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jn97x" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="registry-server" containerID="cri-o://77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b" gracePeriod=2 Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.133598 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.276312 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities\") pod \"48125534-76d4-4d10-982c-bbc34761886f\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.276397 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content\") pod \"48125534-76d4-4d10-982c-bbc34761886f\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.276491 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwlrw\" (UniqueName: \"kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw\") pod \"48125534-76d4-4d10-982c-bbc34761886f\" (UID: \"48125534-76d4-4d10-982c-bbc34761886f\") " Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.277751 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities" (OuterVolumeSpecName: "utilities") pod "48125534-76d4-4d10-982c-bbc34761886f" (UID: "48125534-76d4-4d10-982c-bbc34761886f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.283708 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw" (OuterVolumeSpecName: "kube-api-access-qwlrw") pod "48125534-76d4-4d10-982c-bbc34761886f" (UID: "48125534-76d4-4d10-982c-bbc34761886f"). InnerVolumeSpecName "kube-api-access-qwlrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.335625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48125534-76d4-4d10-982c-bbc34761886f" (UID: "48125534-76d4-4d10-982c-bbc34761886f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.378999 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.379243 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48125534-76d4-4d10-982c-bbc34761886f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.379255 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwlrw\" (UniqueName: \"kubernetes.io/projected/48125534-76d4-4d10-982c-bbc34761886f-kube-api-access-qwlrw\") on node \"crc\" DevicePath \"\"" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.648787 4959 generic.go:334] "Generic (PLEG): container finished" podID="48125534-76d4-4d10-982c-bbc34761886f" containerID="77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b" exitCode=0 Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.648862 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerDied","Data":"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b"} Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.648896 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jn97x" event={"ID":"48125534-76d4-4d10-982c-bbc34761886f","Type":"ContainerDied","Data":"bc9b2f3900ecaf20b3b9dd49cffbad40acd1704eed5f71e1b4924e6bfbd7c223"} Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.648945 4959 scope.go:117] "RemoveContainer" containerID="77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.649246 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jn97x" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.713525 4959 scope.go:117] "RemoveContainer" containerID="af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.719923 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.727085 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jn97x"] Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.743752 4959 scope.go:117] "RemoveContainer" containerID="e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.793104 4959 scope.go:117] "RemoveContainer" containerID="77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b" Oct 03 15:25:37 crc kubenswrapper[4959]: E1003 15:25:37.793874 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b\": container with ID starting with 77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b not found: ID does not exist" containerID="77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.793926 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b"} err="failed to get container status \"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b\": rpc error: code = NotFound desc = could not find container \"77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b\": container with ID starting with 77aedbde24d0556f26a5655e00a0ef4628d5a4d797bdedefd7ebc0debbd1994b not found: ID does not exist" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.793957 4959 scope.go:117] "RemoveContainer" containerID="af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992" Oct 03 15:25:37 crc kubenswrapper[4959]: E1003 15:25:37.794519 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992\": container with ID starting with af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992 not found: ID does not exist" containerID="af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.794589 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992"} err="failed to get container status \"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992\": rpc error: code = NotFound desc = could not find container \"af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992\": container with ID starting with af41ba7952d10731442f342a24173a45a1e497aa8497fd85be14c9abbd27e992 not found: ID does not exist" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.794622 4959 scope.go:117] "RemoveContainer" containerID="e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed" Oct 03 15:25:37 crc kubenswrapper[4959]: E1003 15:25:37.795026 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed\": container with ID starting with e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed not found: ID does not exist" containerID="e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed" Oct 03 15:25:37 crc kubenswrapper[4959]: I1003 15:25:37.795059 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed"} err="failed to get container status \"e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed\": rpc error: code = NotFound desc = could not find container \"e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed\": container with ID starting with e983a1d41003fb6104f5635e0170048383a7834864e57a29757dd052b7e827ed not found: ID does not exist" Oct 03 15:25:39 crc kubenswrapper[4959]: I1003 15:25:39.709849 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48125534-76d4-4d10-982c-bbc34761886f" path="/var/lib/kubelet/pods/48125534-76d4-4d10-982c-bbc34761886f/volumes" Oct 03 15:25:41 crc kubenswrapper[4959]: I1003 15:25:41.055019 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-d30d-account-create-cfnvj"] Oct 03 15:25:41 crc kubenswrapper[4959]: I1003 15:25:41.080437 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-d30d-account-create-cfnvj"] Oct 03 15:25:41 crc kubenswrapper[4959]: I1003 15:25:41.706775 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebce59c6-74ef-431c-8685-ee7eaf0c319a" path="/var/lib/kubelet/pods/ebce59c6-74ef-431c-8685-ee7eaf0c319a/volumes" Oct 03 15:25:54 crc kubenswrapper[4959]: I1003 15:25:54.058917 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-zxwhz"] Oct 03 15:25:54 crc kubenswrapper[4959]: I1003 15:25:54.075429 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-zxwhz"] Oct 03 15:25:55 crc kubenswrapper[4959]: I1003 15:25:55.721738 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="082b2d7d-c1c4-4916-bcd2-55427433e41b" path="/var/lib/kubelet/pods/082b2d7d-c1c4-4916-bcd2-55427433e41b/volumes" Oct 03 15:26:10 crc kubenswrapper[4959]: I1003 15:26:10.943074 4959 scope.go:117] "RemoveContainer" containerID="041b30bde2e90f807045461e862856e8b2143bdf81845e0ab042fe9b8d90f473" Oct 03 15:26:10 crc kubenswrapper[4959]: I1003 15:26:10.992707 4959 scope.go:117] "RemoveContainer" containerID="de08e4d4afcd1d13c69f2957aa766f09fa17fca5a655c50601d1d11b54cefed4" Oct 03 15:26:11 crc kubenswrapper[4959]: I1003 15:26:11.048679 4959 scope.go:117] "RemoveContainer" containerID="84f60ef4f074f7ee031d1ec3c5a017b68b367b8281bee39f60e53e8143cf296e" Oct 03 15:27:06 crc kubenswrapper[4959]: I1003 15:27:06.044427 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:27:06 crc kubenswrapper[4959]: I1003 15:27:06.045173 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:27:36 crc kubenswrapper[4959]: I1003 15:27:36.044707 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:27:36 crc kubenswrapper[4959]: I1003 15:27:36.045151 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:27:56 crc kubenswrapper[4959]: I1003 15:27:56.067634 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-w5d68"] Oct 03 15:27:56 crc kubenswrapper[4959]: I1003 15:27:56.080906 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-w5d68"] Oct 03 15:27:57 crc kubenswrapper[4959]: I1003 15:27:57.701301 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d412c3ef-82c9-46f5-a16c-8765c5f8d8c7" path="/var/lib/kubelet/pods/d412c3ef-82c9-46f5-a16c-8765c5f8d8c7/volumes" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.044133 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.044870 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.044937 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.046117 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.046256 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" gracePeriod=600 Oct 03 15:28:06 crc kubenswrapper[4959]: E1003 15:28:06.166597 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.316585 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" exitCode=0 Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.316665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431"} Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.316735 4959 scope.go:117] "RemoveContainer" containerID="6c7ab24313335440239d2babd17538da0b71a93e969d860ac50c9ee1a2c6ff0b" Oct 03 15:28:06 crc kubenswrapper[4959]: I1003 15:28:06.317584 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:28:06 crc kubenswrapper[4959]: E1003 15:28:06.317915 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:28:07 crc kubenswrapper[4959]: I1003 15:28:07.025810 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-54f8-account-create-8kj78"] Oct 03 15:28:07 crc kubenswrapper[4959]: I1003 15:28:07.034896 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-54f8-account-create-8kj78"] Oct 03 15:28:07 crc kubenswrapper[4959]: I1003 15:28:07.701197 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea38a4f4-157e-442d-b56c-3026346f2849" path="/var/lib/kubelet/pods/ea38a4f4-157e-442d-b56c-3026346f2849/volumes" Oct 03 15:28:11 crc kubenswrapper[4959]: I1003 15:28:11.199373 4959 scope.go:117] "RemoveContainer" containerID="85c85355801d0989a8ca4ebe35ef7e490087c1512a3d43eb9b85a2079eec7da8" Oct 03 15:28:11 crc kubenswrapper[4959]: I1003 15:28:11.236523 4959 scope.go:117] "RemoveContainer" containerID="bc327590ada3a3adef141d2f334bde72c0e992fa75b81b7eba2a57acf85a81da" Oct 03 15:28:19 crc kubenswrapper[4959]: I1003 15:28:19.686761 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:28:19 crc kubenswrapper[4959]: E1003 15:28:19.687943 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:28:23 crc kubenswrapper[4959]: I1003 15:28:23.053136 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-wqxdl"] Oct 03 15:28:23 crc kubenswrapper[4959]: I1003 15:28:23.064570 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-wqxdl"] Oct 03 15:28:23 crc kubenswrapper[4959]: I1003 15:28:23.705976 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd" path="/var/lib/kubelet/pods/0bba5401-d9a5-46a0-8bca-f36dcbc5ecbd/volumes" Oct 03 15:28:30 crc kubenswrapper[4959]: I1003 15:28:30.686758 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:28:30 crc kubenswrapper[4959]: E1003 15:28:30.689241 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:28:45 crc kubenswrapper[4959]: I1003 15:28:45.686570 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:28:45 crc kubenswrapper[4959]: E1003 15:28:45.688083 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:28:47 crc kubenswrapper[4959]: I1003 15:28:47.026969 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-cpk86"] Oct 03 15:28:47 crc kubenswrapper[4959]: I1003 15:28:47.038498 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-cpk86"] Oct 03 15:28:47 crc kubenswrapper[4959]: I1003 15:28:47.703930 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1910f143-5bca-4d35-a37f-82797c1bfdee" path="/var/lib/kubelet/pods/1910f143-5bca-4d35-a37f-82797c1bfdee/volumes" Oct 03 15:28:56 crc kubenswrapper[4959]: I1003 15:28:56.046394 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-94da-account-create-m89k9"] Oct 03 15:28:56 crc kubenswrapper[4959]: I1003 15:28:56.059571 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-94da-account-create-m89k9"] Oct 03 15:28:57 crc kubenswrapper[4959]: I1003 15:28:57.703882 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4e59dfb-dce5-47d2-a21e-dc87b6fc383e" path="/var/lib/kubelet/pods/f4e59dfb-dce5-47d2-a21e-dc87b6fc383e/volumes" Oct 03 15:28:59 crc kubenswrapper[4959]: I1003 15:28:59.686500 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:28:59 crc kubenswrapper[4959]: E1003 15:28:59.689880 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:29:08 crc kubenswrapper[4959]: I1003 15:29:08.055852 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-mfcn8"] Oct 03 15:29:08 crc kubenswrapper[4959]: I1003 15:29:08.063701 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-mfcn8"] Oct 03 15:29:09 crc kubenswrapper[4959]: I1003 15:29:09.710451 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7153a88b-bb05-46b9-8df2-7dff0620e8ba" path="/var/lib/kubelet/pods/7153a88b-bb05-46b9-8df2-7dff0620e8ba/volumes" Oct 03 15:29:11 crc kubenswrapper[4959]: I1003 15:29:11.346326 4959 scope.go:117] "RemoveContainer" containerID="460bdb5d9473ef24810f468a9b7a6aea0c4eb1862bdc0e90c1441f8796073180" Oct 03 15:29:11 crc kubenswrapper[4959]: I1003 15:29:11.391679 4959 scope.go:117] "RemoveContainer" containerID="77743be64b83e19ee0316e2b9c15c3b54ad6cc15e7b78b53ee78c50bcaf80b6d" Oct 03 15:29:11 crc kubenswrapper[4959]: I1003 15:29:11.449715 4959 scope.go:117] "RemoveContainer" containerID="cdb77196f0e9c931a9d3d986c1466a0e505df53d2458e5bb3e82a80a8e3e9d8c" Oct 03 15:29:11 crc kubenswrapper[4959]: I1003 15:29:11.519784 4959 scope.go:117] "RemoveContainer" containerID="c9629f524e34e5d89c359b8d21f736da0e2eae4c7298f5171deb2924ed252f53" Oct 03 15:29:14 crc kubenswrapper[4959]: I1003 15:29:14.686675 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:29:14 crc kubenswrapper[4959]: E1003 15:29:14.687660 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:29:26 crc kubenswrapper[4959]: I1003 15:29:26.686751 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:29:26 crc kubenswrapper[4959]: E1003 15:29:26.688083 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:29:37 crc kubenswrapper[4959]: I1003 15:29:37.693935 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:29:37 crc kubenswrapper[4959]: E1003 15:29:37.694972 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:29:51 crc kubenswrapper[4959]: I1003 15:29:51.685955 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:29:51 crc kubenswrapper[4959]: E1003 15:29:51.688927 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.200813 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7"] Oct 03 15:30:00 crc kubenswrapper[4959]: E1003 15:30:00.201807 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="extract-content" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.201821 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="extract-content" Oct 03 15:30:00 crc kubenswrapper[4959]: E1003 15:30:00.201841 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="extract-utilities" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.201847 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="extract-utilities" Oct 03 15:30:00 crc kubenswrapper[4959]: E1003 15:30:00.201867 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="registry-server" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.201873 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="registry-server" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.202077 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="48125534-76d4-4d10-982c-bbc34761886f" containerName="registry-server" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.202877 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.207585 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.209066 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.214232 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7"] Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.351308 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfmt7\" (UniqueName: \"kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.351916 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.352120 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.454824 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.455007 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfmt7\" (UniqueName: \"kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.455356 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.457090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.469184 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.485381 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfmt7\" (UniqueName: \"kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7\") pod \"collect-profiles-29325090-5gzs7\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:00 crc kubenswrapper[4959]: I1003 15:30:00.549027 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:01 crc kubenswrapper[4959]: I1003 15:30:01.060833 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7"] Oct 03 15:30:01 crc kubenswrapper[4959]: I1003 15:30:01.704532 4959 generic.go:334] "Generic (PLEG): container finished" podID="08619763-d2c8-4d95-955f-c9baaa8d3bd3" containerID="290a1b47e9abef7b05a5a4cb3ed4620a82639d0869cac250f7fbc3a8e66a8bf6" exitCode=0 Oct 03 15:30:01 crc kubenswrapper[4959]: I1003 15:30:01.704642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" event={"ID":"08619763-d2c8-4d95-955f-c9baaa8d3bd3","Type":"ContainerDied","Data":"290a1b47e9abef7b05a5a4cb3ed4620a82639d0869cac250f7fbc3a8e66a8bf6"} Oct 03 15:30:01 crc kubenswrapper[4959]: I1003 15:30:01.704822 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" event={"ID":"08619763-d2c8-4d95-955f-c9baaa8d3bd3","Type":"ContainerStarted","Data":"c65ff79f89966b17388490947ee6144b49564d5961c6179d4bf866e0d7c4fc66"} Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.039588 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.117303 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfmt7\" (UniqueName: \"kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7\") pod \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.117379 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume\") pod \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.117411 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume\") pod \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\" (UID: \"08619763-d2c8-4d95-955f-c9baaa8d3bd3\") " Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.118917 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume" (OuterVolumeSpecName: "config-volume") pod "08619763-d2c8-4d95-955f-c9baaa8d3bd3" (UID: "08619763-d2c8-4d95-955f-c9baaa8d3bd3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.123799 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7" (OuterVolumeSpecName: "kube-api-access-pfmt7") pod "08619763-d2c8-4d95-955f-c9baaa8d3bd3" (UID: "08619763-d2c8-4d95-955f-c9baaa8d3bd3"). InnerVolumeSpecName "kube-api-access-pfmt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.123826 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "08619763-d2c8-4d95-955f-c9baaa8d3bd3" (UID: "08619763-d2c8-4d95-955f-c9baaa8d3bd3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.220068 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfmt7\" (UniqueName: \"kubernetes.io/projected/08619763-d2c8-4d95-955f-c9baaa8d3bd3-kube-api-access-pfmt7\") on node \"crc\" DevicePath \"\"" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.220455 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08619763-d2c8-4d95-955f-c9baaa8d3bd3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.220471 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08619763-d2c8-4d95-955f-c9baaa8d3bd3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.685864 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:30:03 crc kubenswrapper[4959]: E1003 15:30:03.686441 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.726171 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" event={"ID":"08619763-d2c8-4d95-955f-c9baaa8d3bd3","Type":"ContainerDied","Data":"c65ff79f89966b17388490947ee6144b49564d5961c6179d4bf866e0d7c4fc66"} Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.726534 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c65ff79f89966b17388490947ee6144b49564d5961c6179d4bf866e0d7c4fc66" Oct 03 15:30:03 crc kubenswrapper[4959]: I1003 15:30:03.726308 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7" Oct 03 15:30:04 crc kubenswrapper[4959]: I1003 15:30:04.126018 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9"] Oct 03 15:30:04 crc kubenswrapper[4959]: I1003 15:30:04.134485 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325045-tmpv9"] Oct 03 15:30:05 crc kubenswrapper[4959]: I1003 15:30:05.709498 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="015f6ed0-adc6-4144-8adf-bc69a828b80f" path="/var/lib/kubelet/pods/015f6ed0-adc6-4144-8adf-bc69a828b80f/volumes" Oct 03 15:30:11 crc kubenswrapper[4959]: I1003 15:30:11.698993 4959 scope.go:117] "RemoveContainer" containerID="0ea57b1be2e4df0b5a9856387051928c05c85db2b6a2eb28bccc325e9f608fe2" Oct 03 15:30:18 crc kubenswrapper[4959]: I1003 15:30:18.686623 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:30:18 crc kubenswrapper[4959]: E1003 15:30:18.687858 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:30:29 crc kubenswrapper[4959]: I1003 15:30:29.685336 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:30:29 crc kubenswrapper[4959]: E1003 15:30:29.686238 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:30:40 crc kubenswrapper[4959]: I1003 15:30:40.686327 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:30:40 crc kubenswrapper[4959]: E1003 15:30:40.687761 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:30:53 crc kubenswrapper[4959]: I1003 15:30:53.685652 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:30:53 crc kubenswrapper[4959]: E1003 15:30:53.686468 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:31:06 crc kubenswrapper[4959]: I1003 15:31:06.686692 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:31:06 crc kubenswrapper[4959]: E1003 15:31:06.687639 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:31:17 crc kubenswrapper[4959]: I1003 15:31:17.692464 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:31:17 crc kubenswrapper[4959]: E1003 15:31:17.693462 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:31:31 crc kubenswrapper[4959]: I1003 15:31:31.701488 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:31:31 crc kubenswrapper[4959]: E1003 15:31:31.702422 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:31:32 crc kubenswrapper[4959]: I1003 15:31:32.818062 4959 generic.go:334] "Generic (PLEG): container finished" podID="034c6ab7-82c8-4922-ae56-e1862f88d39e" containerID="2b64463677badd0f46938591beb1d06e335cb4e1e9d2426422732b330a23fd46" exitCode=0 Oct 03 15:31:32 crc kubenswrapper[4959]: I1003 15:31:32.818351 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" event={"ID":"034c6ab7-82c8-4922-ae56-e1862f88d39e","Type":"ContainerDied","Data":"2b64463677badd0f46938591beb1d06e335cb4e1e9d2426422732b330a23fd46"} Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.328749 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.447095 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph\") pod \"034c6ab7-82c8-4922-ae56-e1862f88d39e\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.447351 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key\") pod \"034c6ab7-82c8-4922-ae56-e1862f88d39e\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.447444 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle\") pod \"034c6ab7-82c8-4922-ae56-e1862f88d39e\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.447563 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl5nm\" (UniqueName: \"kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm\") pod \"034c6ab7-82c8-4922-ae56-e1862f88d39e\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.447825 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory\") pod \"034c6ab7-82c8-4922-ae56-e1862f88d39e\" (UID: \"034c6ab7-82c8-4922-ae56-e1862f88d39e\") " Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.453119 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm" (OuterVolumeSpecName: "kube-api-access-kl5nm") pod "034c6ab7-82c8-4922-ae56-e1862f88d39e" (UID: "034c6ab7-82c8-4922-ae56-e1862f88d39e"). InnerVolumeSpecName "kube-api-access-kl5nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.454678 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph" (OuterVolumeSpecName: "ceph") pod "034c6ab7-82c8-4922-ae56-e1862f88d39e" (UID: "034c6ab7-82c8-4922-ae56-e1862f88d39e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.454707 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "034c6ab7-82c8-4922-ae56-e1862f88d39e" (UID: "034c6ab7-82c8-4922-ae56-e1862f88d39e"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.483775 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "034c6ab7-82c8-4922-ae56-e1862f88d39e" (UID: "034c6ab7-82c8-4922-ae56-e1862f88d39e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.510051 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory" (OuterVolumeSpecName: "inventory") pod "034c6ab7-82c8-4922-ae56-e1862f88d39e" (UID: "034c6ab7-82c8-4922-ae56-e1862f88d39e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.550094 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl5nm\" (UniqueName: \"kubernetes.io/projected/034c6ab7-82c8-4922-ae56-e1862f88d39e-kube-api-access-kl5nm\") on node \"crc\" DevicePath \"\"" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.550123 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.550132 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.550140 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.550150 4959 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/034c6ab7-82c8-4922-ae56-e1862f88d39e-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.838748 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" event={"ID":"034c6ab7-82c8-4922-ae56-e1862f88d39e","Type":"ContainerDied","Data":"b2c2da22d2f828cd6df5eaf305d4e45dea7551fdb7feb2d98cabe47f7b1ff390"} Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.838799 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g" Oct 03 15:31:34 crc kubenswrapper[4959]: I1003 15:31:34.838822 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2c2da22d2f828cd6df5eaf305d4e45dea7551fdb7feb2d98cabe47f7b1ff390" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.411841 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-6vxnz"] Oct 03 15:31:38 crc kubenswrapper[4959]: E1003 15:31:38.413311 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08619763-d2c8-4d95-955f-c9baaa8d3bd3" containerName="collect-profiles" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.413358 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="08619763-d2c8-4d95-955f-c9baaa8d3bd3" containerName="collect-profiles" Oct 03 15:31:38 crc kubenswrapper[4959]: E1003 15:31:38.413395 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034c6ab7-82c8-4922-ae56-e1862f88d39e" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.413410 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="034c6ab7-82c8-4922-ae56-e1862f88d39e" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.413817 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="08619763-d2c8-4d95-955f-c9baaa8d3bd3" containerName="collect-profiles" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.413845 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="034c6ab7-82c8-4922-ae56-e1862f88d39e" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.415029 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.421374 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.421411 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.423568 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.424993 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.432371 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-6vxnz"] Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.551177 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zgk6\" (UniqueName: \"kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.551355 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.551550 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.551605 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.551670 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.654373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zgk6\" (UniqueName: \"kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.654510 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.654628 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.654673 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.654746 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.662986 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.665989 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.667906 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.671142 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.683343 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zgk6\" (UniqueName: \"kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6\") pod \"bootstrap-openstack-openstack-cell1-6vxnz\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:38 crc kubenswrapper[4959]: I1003 15:31:38.736684 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:31:39 crc kubenswrapper[4959]: I1003 15:31:39.316367 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:31:39 crc kubenswrapper[4959]: I1003 15:31:39.323344 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-6vxnz"] Oct 03 15:31:39 crc kubenswrapper[4959]: I1003 15:31:39.910792 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" event={"ID":"faa34915-61b3-4e8f-a90b-b4d45cace539","Type":"ContainerStarted","Data":"5bb107bb9ee42130f4601fcfce7358f33a0bce13bf6cefbff1a758b4429c5cc8"} Oct 03 15:31:39 crc kubenswrapper[4959]: I1003 15:31:39.911067 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" event={"ID":"faa34915-61b3-4e8f-a90b-b4d45cace539","Type":"ContainerStarted","Data":"fc71be485cb48a62b4f058633c6c536d3697a52da602f582e82f7f01962d9fa1"} Oct 03 15:31:39 crc kubenswrapper[4959]: I1003 15:31:39.934848 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" podStartSLOduration=1.745969866 podStartE2EDuration="1.934818842s" podCreationTimestamp="2025-10-03 15:31:38 +0000 UTC" firstStartedPulling="2025-10-03 15:31:39.31617808 +0000 UTC m=+7268.519521497" lastFinishedPulling="2025-10-03 15:31:39.505027056 +0000 UTC m=+7268.708370473" observedRunningTime="2025-10-03 15:31:39.926500359 +0000 UTC m=+7269.129843816" watchObservedRunningTime="2025-10-03 15:31:39.934818842 +0000 UTC m=+7269.138162299" Oct 03 15:31:44 crc kubenswrapper[4959]: I1003 15:31:44.685881 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:31:44 crc kubenswrapper[4959]: E1003 15:31:44.686745 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:31:56 crc kubenswrapper[4959]: I1003 15:31:56.686987 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:31:56 crc kubenswrapper[4959]: E1003 15:31:56.687773 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:32:08 crc kubenswrapper[4959]: I1003 15:32:08.685893 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:32:08 crc kubenswrapper[4959]: E1003 15:32:08.687006 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:32:19 crc kubenswrapper[4959]: I1003 15:32:19.686288 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:32:19 crc kubenswrapper[4959]: E1003 15:32:19.688995 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:32:34 crc kubenswrapper[4959]: I1003 15:32:34.687068 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:32:34 crc kubenswrapper[4959]: E1003 15:32:34.688908 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:32:49 crc kubenswrapper[4959]: I1003 15:32:49.686054 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:32:49 crc kubenswrapper[4959]: E1003 15:32:49.687131 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:33:00 crc kubenswrapper[4959]: I1003 15:33:00.685905 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:33:00 crc kubenswrapper[4959]: E1003 15:33:00.686634 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:33:13 crc kubenswrapper[4959]: I1003 15:33:13.685973 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:33:14 crc kubenswrapper[4959]: I1003 15:33:14.098409 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030"} Oct 03 15:33:26 crc kubenswrapper[4959]: I1003 15:33:26.780398 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 03 15:33:31 crc kubenswrapper[4959]: I1003 15:33:31.780334 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 03 15:33:31 crc kubenswrapper[4959]: I1003 15:33:31.779175 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Oct 03 15:33:32 crc kubenswrapper[4959]: I1003 15:33:32.774877 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-central-agent" probeResult="failure" output=< Oct 03 15:33:32 crc kubenswrapper[4959]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 03 15:33:32 crc kubenswrapper[4959]: > Oct 03 15:33:32 crc kubenswrapper[4959]: I1003 15:33:32.775213 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 03 15:33:32 crc kubenswrapper[4959]: I1003 15:33:32.776067 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"8a8b567a92b733c71dbbe9b9b83175fe00dbd21de71a2cfc19d1cbdab50d31cf"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Oct 03 15:33:32 crc kubenswrapper[4959]: I1003 15:33:32.776154 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-central-agent" containerID="cri-o://8a8b567a92b733c71dbbe9b9b83175fe00dbd21de71a2cfc19d1cbdab50d31cf" gracePeriod=30 Oct 03 15:33:33 crc kubenswrapper[4959]: I1003 15:33:33.320677 4959 generic.go:334] "Generic (PLEG): container finished" podID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerID="8a8b567a92b733c71dbbe9b9b83175fe00dbd21de71a2cfc19d1cbdab50d31cf" exitCode=0 Oct 03 15:33:33 crc kubenswrapper[4959]: I1003 15:33:33.320716 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerDied","Data":"8a8b567a92b733c71dbbe9b9b83175fe00dbd21de71a2cfc19d1cbdab50d31cf"} Oct 03 15:33:34 crc kubenswrapper[4959]: I1003 15:33:34.337766 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"7a8cf30ec6efa5e843422bce60e5d75267f9eb38e3583e9c6bda001c2af8b6be"} Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.009859 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.013373 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.027858 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.072422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.072627 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkzxc\" (UniqueName: \"kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.072665 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.175216 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.175354 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkzxc\" (UniqueName: \"kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.175383 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.175832 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.176320 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.207015 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkzxc\" (UniqueName: \"kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc\") pod \"community-operators-bdndw\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.339468 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:44 crc kubenswrapper[4959]: I1003 15:33:44.975418 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:45 crc kubenswrapper[4959]: I1003 15:33:45.492966 4959 generic.go:334] "Generic (PLEG): container finished" podID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerID="8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35" exitCode=0 Oct 03 15:33:45 crc kubenswrapper[4959]: I1003 15:33:45.493032 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerDied","Data":"8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35"} Oct 03 15:33:45 crc kubenswrapper[4959]: I1003 15:33:45.493358 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerStarted","Data":"805b73cda743e2e152c948c814416fdb59fae0f0265e8273a274ce4c25e94525"} Oct 03 15:33:47 crc kubenswrapper[4959]: I1003 15:33:47.518411 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerStarted","Data":"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c"} Oct 03 15:33:48 crc kubenswrapper[4959]: I1003 15:33:48.528150 4959 generic.go:334] "Generic (PLEG): container finished" podID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerID="fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c" exitCode=0 Oct 03 15:33:48 crc kubenswrapper[4959]: I1003 15:33:48.528299 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerDied","Data":"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c"} Oct 03 15:33:49 crc kubenswrapper[4959]: I1003 15:33:49.543942 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerStarted","Data":"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c"} Oct 03 15:33:49 crc kubenswrapper[4959]: I1003 15:33:49.566057 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bdndw" podStartSLOduration=3.131123934 podStartE2EDuration="6.56604133s" podCreationTimestamp="2025-10-03 15:33:43 +0000 UTC" firstStartedPulling="2025-10-03 15:33:45.496533772 +0000 UTC m=+7394.699877209" lastFinishedPulling="2025-10-03 15:33:48.931451188 +0000 UTC m=+7398.134794605" observedRunningTime="2025-10-03 15:33:49.564792509 +0000 UTC m=+7398.768135946" watchObservedRunningTime="2025-10-03 15:33:49.56604133 +0000 UTC m=+7398.769384737" Oct 03 15:33:54 crc kubenswrapper[4959]: I1003 15:33:54.339721 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:54 crc kubenswrapper[4959]: I1003 15:33:54.340406 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:54 crc kubenswrapper[4959]: I1003 15:33:54.418842 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:54 crc kubenswrapper[4959]: I1003 15:33:54.655146 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:54 crc kubenswrapper[4959]: I1003 15:33:54.710322 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:56 crc kubenswrapper[4959]: I1003 15:33:56.614951 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bdndw" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="registry-server" containerID="cri-o://97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c" gracePeriod=2 Oct 03 15:33:56 crc kubenswrapper[4959]: I1003 15:33:56.622923 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-notification-agent" probeResult="failure" output=< Oct 03 15:33:56 crc kubenswrapper[4959]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 03 15:33:56 crc kubenswrapper[4959]: > Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.252147 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.396029 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities\") pod \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.396138 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content\") pod \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.396382 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkzxc\" (UniqueName: \"kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc\") pod \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\" (UID: \"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3\") " Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.396848 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities" (OuterVolumeSpecName: "utilities") pod "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" (UID: "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.409515 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc" (OuterVolumeSpecName: "kube-api-access-wkzxc") pod "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" (UID: "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3"). InnerVolumeSpecName "kube-api-access-wkzxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.454968 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" (UID: "a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.509531 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkzxc\" (UniqueName: \"kubernetes.io/projected/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-kube-api-access-wkzxc\") on node \"crc\" DevicePath \"\"" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.509580 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.509599 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.625260 4959 generic.go:334] "Generic (PLEG): container finished" podID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerID="97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c" exitCode=0 Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.626094 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerDied","Data":"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c"} Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.626473 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bdndw" event={"ID":"a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3","Type":"ContainerDied","Data":"805b73cda743e2e152c948c814416fdb59fae0f0265e8273a274ce4c25e94525"} Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.626168 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bdndw" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.626623 4959 scope.go:117] "RemoveContainer" containerID="97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.670966 4959 scope.go:117] "RemoveContainer" containerID="fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.674917 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.684401 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bdndw"] Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.701525 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" path="/var/lib/kubelet/pods/a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3/volumes" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.713526 4959 scope.go:117] "RemoveContainer" containerID="8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.762019 4959 scope.go:117] "RemoveContainer" containerID="97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c" Oct 03 15:33:57 crc kubenswrapper[4959]: E1003 15:33:57.762618 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c\": container with ID starting with 97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c not found: ID does not exist" containerID="97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.762663 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c"} err="failed to get container status \"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c\": rpc error: code = NotFound desc = could not find container \"97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c\": container with ID starting with 97ba4f3d122982bb81afccbfc8fa96b59f3fd50da928fdd106cacfc446b37b1c not found: ID does not exist" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.762725 4959 scope.go:117] "RemoveContainer" containerID="fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c" Oct 03 15:33:57 crc kubenswrapper[4959]: E1003 15:33:57.763288 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c\": container with ID starting with fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c not found: ID does not exist" containerID="fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.763323 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c"} err="failed to get container status \"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c\": rpc error: code = NotFound desc = could not find container \"fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c\": container with ID starting with fca8e5fda0e71ce73915923bf7c2d013a8c6d1b42f1f7962d192d7e423fe061c not found: ID does not exist" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.763346 4959 scope.go:117] "RemoveContainer" containerID="8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35" Oct 03 15:33:57 crc kubenswrapper[4959]: E1003 15:33:57.763718 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35\": container with ID starting with 8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35 not found: ID does not exist" containerID="8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35" Oct 03 15:33:57 crc kubenswrapper[4959]: I1003 15:33:57.763746 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35"} err="failed to get container status \"8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35\": rpc error: code = NotFound desc = could not find container \"8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35\": container with ID starting with 8ad2c527f5ee5444e57996b18d43439e7b54688a955609b286bc775b764b6f35 not found: ID does not exist" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.343279 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:19 crc kubenswrapper[4959]: E1003 15:34:19.344852 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="extract-utilities" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.344887 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="extract-utilities" Oct 03 15:34:19 crc kubenswrapper[4959]: E1003 15:34:19.344989 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="registry-server" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.345007 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="registry-server" Oct 03 15:34:19 crc kubenswrapper[4959]: E1003 15:34:19.345924 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="extract-content" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.345975 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="extract-content" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.346402 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73bfb98-4b9c-42f4-97cf-4dfa12dc9bf3" containerName="registry-server" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.349597 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.364819 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.421595 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.421641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.421727 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv4sj\" (UniqueName: \"kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.524085 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv4sj\" (UniqueName: \"kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.524245 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.524275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.525112 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.525139 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.560075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv4sj\" (UniqueName: \"kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj\") pod \"redhat-operators-bbv7z\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:19 crc kubenswrapper[4959]: I1003 15:34:19.712670 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:20 crc kubenswrapper[4959]: I1003 15:34:20.230554 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:20 crc kubenswrapper[4959]: I1003 15:34:20.925235 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerDied","Data":"c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d"} Oct 03 15:34:20 crc kubenswrapper[4959]: I1003 15:34:20.925155 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerID="c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d" exitCode=0 Oct 03 15:34:20 crc kubenswrapper[4959]: I1003 15:34:20.925665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerStarted","Data":"879f92ad1de7c5fd918cde0e41a25de22a12ddfdc3ea123800011951f3394d23"} Oct 03 15:34:22 crc kubenswrapper[4959]: I1003 15:34:22.950013 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerStarted","Data":"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e"} Oct 03 15:34:25 crc kubenswrapper[4959]: I1003 15:34:25.991764 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerID="d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e" exitCode=0 Oct 03 15:34:25 crc kubenswrapper[4959]: I1003 15:34:25.991977 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerDied","Data":"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e"} Oct 03 15:34:26 crc kubenswrapper[4959]: I1003 15:34:26.669909 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-notification-agent" probeResult="failure" output=< Oct 03 15:34:26 crc kubenswrapper[4959]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 03 15:34:26 crc kubenswrapper[4959]: > Oct 03 15:34:26 crc kubenswrapper[4959]: I1003 15:34:26.670044 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 03 15:34:26 crc kubenswrapper[4959]: I1003 15:34:26.671565 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-notification-agent" containerStatusID={"Type":"cri-o","ID":"29097b82b5456dc9ca66762a31428a87e3c8f5c6856b29170584e4d9f8be1150"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-notification-agent failed liveness probe, will be restarted" Oct 03 15:34:26 crc kubenswrapper[4959]: I1003 15:34:26.671679 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerName="ceilometer-notification-agent" containerID="cri-o://29097b82b5456dc9ca66762a31428a87e3c8f5c6856b29170584e4d9f8be1150" gracePeriod=30 Oct 03 15:34:27 crc kubenswrapper[4959]: I1003 15:34:27.010187 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerStarted","Data":"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee"} Oct 03 15:34:27 crc kubenswrapper[4959]: I1003 15:34:27.052553 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bbv7z" podStartSLOduration=2.287632407 podStartE2EDuration="8.052523258s" podCreationTimestamp="2025-10-03 15:34:19 +0000 UTC" firstStartedPulling="2025-10-03 15:34:20.927064324 +0000 UTC m=+7430.130407771" lastFinishedPulling="2025-10-03 15:34:26.691955165 +0000 UTC m=+7435.895298622" observedRunningTime="2025-10-03 15:34:27.034508177 +0000 UTC m=+7436.237851634" watchObservedRunningTime="2025-10-03 15:34:27.052523258 +0000 UTC m=+7436.255866705" Oct 03 15:34:29 crc kubenswrapper[4959]: I1003 15:34:29.036591 4959 generic.go:334] "Generic (PLEG): container finished" podID="5922a823-0599-4108-a7ca-d7eb68de6c2d" containerID="29097b82b5456dc9ca66762a31428a87e3c8f5c6856b29170584e4d9f8be1150" exitCode=0 Oct 03 15:34:29 crc kubenswrapper[4959]: I1003 15:34:29.036746 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerDied","Data":"29097b82b5456dc9ca66762a31428a87e3c8f5c6856b29170584e4d9f8be1150"} Oct 03 15:34:29 crc kubenswrapper[4959]: I1003 15:34:29.712799 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:29 crc kubenswrapper[4959]: I1003 15:34:29.713100 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:30 crc kubenswrapper[4959]: I1003 15:34:30.050352 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5922a823-0599-4108-a7ca-d7eb68de6c2d","Type":"ContainerStarted","Data":"532764a16b2540b87273f1fda84be9aeeb1fa916dca13d84b02173b686f242bc"} Oct 03 15:34:30 crc kubenswrapper[4959]: I1003 15:34:30.788510 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bbv7z" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="registry-server" probeResult="failure" output=< Oct 03 15:34:30 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:34:30 crc kubenswrapper[4959]: > Oct 03 15:34:39 crc kubenswrapper[4959]: I1003 15:34:39.786460 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:39 crc kubenswrapper[4959]: I1003 15:34:39.855457 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:40 crc kubenswrapper[4959]: I1003 15:34:40.033329 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.190974 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bbv7z" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="registry-server" containerID="cri-o://189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee" gracePeriod=2 Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.765159 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.869076 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv4sj\" (UniqueName: \"kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj\") pod \"8d9cb846-16f9-4067-ace2-d5df2c1181de\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.869236 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities\") pod \"8d9cb846-16f9-4067-ace2-d5df2c1181de\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.869433 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") pod \"8d9cb846-16f9-4067-ace2-d5df2c1181de\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.870354 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities" (OuterVolumeSpecName: "utilities") pod "8d9cb846-16f9-4067-ace2-d5df2c1181de" (UID: "8d9cb846-16f9-4067-ace2-d5df2c1181de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.872774 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.875272 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj" (OuterVolumeSpecName: "kube-api-access-dv4sj") pod "8d9cb846-16f9-4067-ace2-d5df2c1181de" (UID: "8d9cb846-16f9-4067-ace2-d5df2c1181de"). InnerVolumeSpecName "kube-api-access-dv4sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.973767 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d9cb846-16f9-4067-ace2-d5df2c1181de" (UID: "8d9cb846-16f9-4067-ace2-d5df2c1181de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.975188 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") pod \"8d9cb846-16f9-4067-ace2-d5df2c1181de\" (UID: \"8d9cb846-16f9-4067-ace2-d5df2c1181de\") " Oct 03 15:34:41 crc kubenswrapper[4959]: W1003 15:34:41.975369 4959 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8d9cb846-16f9-4067-ace2-d5df2c1181de/volumes/kubernetes.io~empty-dir/catalog-content Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.975406 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d9cb846-16f9-4067-ace2-d5df2c1181de" (UID: "8d9cb846-16f9-4067-ace2-d5df2c1181de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.976364 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv4sj\" (UniqueName: \"kubernetes.io/projected/8d9cb846-16f9-4067-ace2-d5df2c1181de-kube-api-access-dv4sj\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:41 crc kubenswrapper[4959]: I1003 15:34:41.976422 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d9cb846-16f9-4067-ace2-d5df2c1181de-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.209281 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerID="189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee" exitCode=0 Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.209347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerDied","Data":"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee"} Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.209389 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bbv7z" event={"ID":"8d9cb846-16f9-4067-ace2-d5df2c1181de","Type":"ContainerDied","Data":"879f92ad1de7c5fd918cde0e41a25de22a12ddfdc3ea123800011951f3394d23"} Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.209407 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bbv7z" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.209434 4959 scope.go:117] "RemoveContainer" containerID="189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.249127 4959 scope.go:117] "RemoveContainer" containerID="d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.303814 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.305774 4959 scope.go:117] "RemoveContainer" containerID="c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.321634 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bbv7z"] Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.375182 4959 scope.go:117] "RemoveContainer" containerID="189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee" Oct 03 15:34:42 crc kubenswrapper[4959]: E1003 15:34:42.376188 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee\": container with ID starting with 189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee not found: ID does not exist" containerID="189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.376285 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee"} err="failed to get container status \"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee\": rpc error: code = NotFound desc = could not find container \"189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee\": container with ID starting with 189baa8928505df9f362f2db6140ce6bce25a85eb1c504370d98eb3b57a155ee not found: ID does not exist" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.376334 4959 scope.go:117] "RemoveContainer" containerID="d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e" Oct 03 15:34:42 crc kubenswrapper[4959]: E1003 15:34:42.376925 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e\": container with ID starting with d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e not found: ID does not exist" containerID="d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.376968 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e"} err="failed to get container status \"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e\": rpc error: code = NotFound desc = could not find container \"d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e\": container with ID starting with d5838959bc2c75411dcf0a6b60a428ab33565530a4873162c90734bcabe8304e not found: ID does not exist" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.376995 4959 scope.go:117] "RemoveContainer" containerID="c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d" Oct 03 15:34:42 crc kubenswrapper[4959]: E1003 15:34:42.377743 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d\": container with ID starting with c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d not found: ID does not exist" containerID="c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d" Oct 03 15:34:42 crc kubenswrapper[4959]: I1003 15:34:42.377827 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d"} err="failed to get container status \"c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d\": rpc error: code = NotFound desc = could not find container \"c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d\": container with ID starting with c80c2891968db111e932eb65ce471f0421b30f9d2e7b83fd12505eaf36e4da6d not found: ID does not exist" Oct 03 15:34:43 crc kubenswrapper[4959]: I1003 15:34:43.706112 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" path="/var/lib/kubelet/pods/8d9cb846-16f9-4067-ace2-d5df2c1181de/volumes" Oct 03 15:34:50 crc kubenswrapper[4959]: I1003 15:34:50.310268 4959 generic.go:334] "Generic (PLEG): container finished" podID="faa34915-61b3-4e8f-a90b-b4d45cace539" containerID="5bb107bb9ee42130f4601fcfce7358f33a0bce13bf6cefbff1a758b4429c5cc8" exitCode=0 Oct 03 15:34:50 crc kubenswrapper[4959]: I1003 15:34:50.310383 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" event={"ID":"faa34915-61b3-4e8f-a90b-b4d45cace539","Type":"ContainerDied","Data":"5bb107bb9ee42130f4601fcfce7358f33a0bce13bf6cefbff1a758b4429c5cc8"} Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.889955 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.942977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zgk6\" (UniqueName: \"kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6\") pod \"faa34915-61b3-4e8f-a90b-b4d45cace539\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.943318 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle\") pod \"faa34915-61b3-4e8f-a90b-b4d45cace539\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.943872 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph\") pod \"faa34915-61b3-4e8f-a90b-b4d45cace539\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.944212 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key\") pod \"faa34915-61b3-4e8f-a90b-b4d45cace539\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.944955 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory\") pod \"faa34915-61b3-4e8f-a90b-b4d45cace539\" (UID: \"faa34915-61b3-4e8f-a90b-b4d45cace539\") " Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.951758 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6" (OuterVolumeSpecName: "kube-api-access-8zgk6") pod "faa34915-61b3-4e8f-a90b-b4d45cace539" (UID: "faa34915-61b3-4e8f-a90b-b4d45cace539"). InnerVolumeSpecName "kube-api-access-8zgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.952140 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph" (OuterVolumeSpecName: "ceph") pod "faa34915-61b3-4e8f-a90b-b4d45cace539" (UID: "faa34915-61b3-4e8f-a90b-b4d45cace539"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.952611 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.952746 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zgk6\" (UniqueName: \"kubernetes.io/projected/faa34915-61b3-4e8f-a90b-b4d45cace539-kube-api-access-8zgk6\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.959006 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "faa34915-61b3-4e8f-a90b-b4d45cace539" (UID: "faa34915-61b3-4e8f-a90b-b4d45cace539"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.983941 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory" (OuterVolumeSpecName: "inventory") pod "faa34915-61b3-4e8f-a90b-b4d45cace539" (UID: "faa34915-61b3-4e8f-a90b-b4d45cace539"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:34:51 crc kubenswrapper[4959]: I1003 15:34:51.994253 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "faa34915-61b3-4e8f-a90b-b4d45cace539" (UID: "faa34915-61b3-4e8f-a90b-b4d45cace539"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.055042 4959 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.055740 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.055900 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faa34915-61b3-4e8f-a90b-b4d45cace539-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.342825 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" event={"ID":"faa34915-61b3-4e8f-a90b-b4d45cace539","Type":"ContainerDied","Data":"fc71be485cb48a62b4f058633c6c536d3697a52da602f582e82f7f01962d9fa1"} Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.342884 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc71be485cb48a62b4f058633c6c536d3697a52da602f582e82f7f01962d9fa1" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.342963 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-6vxnz" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.461038 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-jlh5w"] Oct 03 15:34:52 crc kubenswrapper[4959]: E1003 15:34:52.461881 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa34915-61b3-4e8f-a90b-b4d45cace539" containerName="bootstrap-openstack-openstack-cell1" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.461996 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa34915-61b3-4e8f-a90b-b4d45cace539" containerName="bootstrap-openstack-openstack-cell1" Oct 03 15:34:52 crc kubenswrapper[4959]: E1003 15:34:52.462110 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="registry-server" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.462214 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="registry-server" Oct 03 15:34:52 crc kubenswrapper[4959]: E1003 15:34:52.462320 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="extract-utilities" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.462404 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="extract-utilities" Oct 03 15:34:52 crc kubenswrapper[4959]: E1003 15:34:52.462508 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="extract-content" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.462611 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="extract-content" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.462992 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9cb846-16f9-4067-ace2-d5df2c1181de" containerName="registry-server" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.463093 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa34915-61b3-4e8f-a90b-b4d45cace539" containerName="bootstrap-openstack-openstack-cell1" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.464097 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.470976 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.471244 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.471898 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.472106 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.484987 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-jlh5w"] Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.568291 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.568419 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.568540 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.568650 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsslj\" (UniqueName: \"kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.670004 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsslj\" (UniqueName: \"kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.670450 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.670600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.670703 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.674880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.679308 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.684927 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.703210 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsslj\" (UniqueName: \"kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj\") pod \"download-cache-openstack-openstack-cell1-jlh5w\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:52 crc kubenswrapper[4959]: I1003 15:34:52.801330 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:34:53 crc kubenswrapper[4959]: I1003 15:34:53.383724 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-jlh5w"] Oct 03 15:34:54 crc kubenswrapper[4959]: I1003 15:34:54.363641 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" event={"ID":"7b050197-797a-4c1d-8e40-530801ae7bd3","Type":"ContainerStarted","Data":"c0810cb4af6de8bb822b949a03b645d2be412d3da9f9ded328311c632252c16f"} Oct 03 15:34:54 crc kubenswrapper[4959]: I1003 15:34:54.363991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" event={"ID":"7b050197-797a-4c1d-8e40-530801ae7bd3","Type":"ContainerStarted","Data":"e22893defb781f84374c141507db6dde65d7638f44bba3ef7e32b844732078ff"} Oct 03 15:34:54 crc kubenswrapper[4959]: I1003 15:34:54.397893 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" podStartSLOduration=2.200690011 podStartE2EDuration="2.397874981s" podCreationTimestamp="2025-10-03 15:34:52 +0000 UTC" firstStartedPulling="2025-10-03 15:34:53.392734391 +0000 UTC m=+7462.596077808" lastFinishedPulling="2025-10-03 15:34:53.589919371 +0000 UTC m=+7462.793262778" observedRunningTime="2025-10-03 15:34:54.389753962 +0000 UTC m=+7463.593097399" watchObservedRunningTime="2025-10-03 15:34:54.397874981 +0000 UTC m=+7463.601218398" Oct 03 15:35:36 crc kubenswrapper[4959]: I1003 15:35:36.043827 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:35:36 crc kubenswrapper[4959]: I1003 15:35:36.044469 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.924229 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.926901 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.952063 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.970641 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5cbs\" (UniqueName: \"kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.970717 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:01 crc kubenswrapper[4959]: I1003 15:36:01.970841 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.072967 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.073057 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5cbs\" (UniqueName: \"kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.073126 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.073622 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.073661 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.099761 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5cbs\" (UniqueName: \"kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs\") pod \"certified-operators-v2rvj\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.250868 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:02 crc kubenswrapper[4959]: I1003 15:36:02.762925 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:03 crc kubenswrapper[4959]: I1003 15:36:03.196958 4959 generic.go:334] "Generic (PLEG): container finished" podID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerID="80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd" exitCode=0 Oct 03 15:36:03 crc kubenswrapper[4959]: I1003 15:36:03.197070 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerDied","Data":"80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd"} Oct 03 15:36:03 crc kubenswrapper[4959]: I1003 15:36:03.197312 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerStarted","Data":"3e99df0c8066e14186ff8f5de834a08f5d95d759374f870ee16df0f39a9f37b1"} Oct 03 15:36:05 crc kubenswrapper[4959]: I1003 15:36:05.221529 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerStarted","Data":"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986"} Oct 03 15:36:06 crc kubenswrapper[4959]: I1003 15:36:06.043940 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:36:06 crc kubenswrapper[4959]: I1003 15:36:06.044313 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:36:06 crc kubenswrapper[4959]: I1003 15:36:06.236599 4959 generic.go:334] "Generic (PLEG): container finished" podID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerID="02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986" exitCode=0 Oct 03 15:36:06 crc kubenswrapper[4959]: I1003 15:36:06.236656 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerDied","Data":"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986"} Oct 03 15:36:07 crc kubenswrapper[4959]: I1003 15:36:07.253981 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerStarted","Data":"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348"} Oct 03 15:36:07 crc kubenswrapper[4959]: I1003 15:36:07.291780 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v2rvj" podStartSLOduration=2.8129180270000003 podStartE2EDuration="6.291759535s" podCreationTimestamp="2025-10-03 15:36:01 +0000 UTC" firstStartedPulling="2025-10-03 15:36:03.200456605 +0000 UTC m=+7532.403800062" lastFinishedPulling="2025-10-03 15:36:06.679298143 +0000 UTC m=+7535.882641570" observedRunningTime="2025-10-03 15:36:07.285408219 +0000 UTC m=+7536.488751686" watchObservedRunningTime="2025-10-03 15:36:07.291759535 +0000 UTC m=+7536.495102962" Oct 03 15:36:12 crc kubenswrapper[4959]: I1003 15:36:12.250947 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:12 crc kubenswrapper[4959]: I1003 15:36:12.252843 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:12 crc kubenswrapper[4959]: I1003 15:36:12.309832 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:13 crc kubenswrapper[4959]: I1003 15:36:13.384212 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:13 crc kubenswrapper[4959]: I1003 15:36:13.453014 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.338921 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v2rvj" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="registry-server" containerID="cri-o://5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348" gracePeriod=2 Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.833865 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.990334 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content\") pod \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.990498 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities\") pod \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.990559 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5cbs\" (UniqueName: \"kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs\") pod \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\" (UID: \"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e\") " Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.992296 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities" (OuterVolumeSpecName: "utilities") pod "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" (UID: "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:36:15 crc kubenswrapper[4959]: I1003 15:36:15.998274 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs" (OuterVolumeSpecName: "kube-api-access-t5cbs") pod "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" (UID: "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e"). InnerVolumeSpecName "kube-api-access-t5cbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.056876 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" (UID: "3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.093371 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.093399 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.093410 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5cbs\" (UniqueName: \"kubernetes.io/projected/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e-kube-api-access-t5cbs\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.351669 4959 generic.go:334] "Generic (PLEG): container finished" podID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerID="5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348" exitCode=0 Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.351727 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerDied","Data":"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348"} Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.351991 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v2rvj" event={"ID":"3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e","Type":"ContainerDied","Data":"3e99df0c8066e14186ff8f5de834a08f5d95d759374f870ee16df0f39a9f37b1"} Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.352012 4959 scope.go:117] "RemoveContainer" containerID="5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.351747 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v2rvj" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.379996 4959 scope.go:117] "RemoveContainer" containerID="02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.397964 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.416509 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v2rvj"] Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.424651 4959 scope.go:117] "RemoveContainer" containerID="80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.460450 4959 scope.go:117] "RemoveContainer" containerID="5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348" Oct 03 15:36:16 crc kubenswrapper[4959]: E1003 15:36:16.460955 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348\": container with ID starting with 5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348 not found: ID does not exist" containerID="5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.460997 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348"} err="failed to get container status \"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348\": rpc error: code = NotFound desc = could not find container \"5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348\": container with ID starting with 5fda966a1fbf803fef0618673036e8f1c8c02b71ef34694077e165c58795a348 not found: ID does not exist" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.461025 4959 scope.go:117] "RemoveContainer" containerID="02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986" Oct 03 15:36:16 crc kubenswrapper[4959]: E1003 15:36:16.462292 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986\": container with ID starting with 02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986 not found: ID does not exist" containerID="02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.462344 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986"} err="failed to get container status \"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986\": rpc error: code = NotFound desc = could not find container \"02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986\": container with ID starting with 02a77b639f378507c7b7d16b43d3d15e5757d44c3267c029103f4d13d43af986 not found: ID does not exist" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.462370 4959 scope.go:117] "RemoveContainer" containerID="80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd" Oct 03 15:36:16 crc kubenswrapper[4959]: E1003 15:36:16.462856 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd\": container with ID starting with 80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd not found: ID does not exist" containerID="80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd" Oct 03 15:36:16 crc kubenswrapper[4959]: I1003 15:36:16.462887 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd"} err="failed to get container status \"80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd\": rpc error: code = NotFound desc = could not find container \"80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd\": container with ID starting with 80174b84d21afb77cd403b71c0fe8052d9325a80583d9682eb534b6a5b00c1dd not found: ID does not exist" Oct 03 15:36:17 crc kubenswrapper[4959]: I1003 15:36:17.698573 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" path="/var/lib/kubelet/pods/3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e/volumes" Oct 03 15:36:29 crc kubenswrapper[4959]: I1003 15:36:29.521243 4959 generic.go:334] "Generic (PLEG): container finished" podID="7b050197-797a-4c1d-8e40-530801ae7bd3" containerID="c0810cb4af6de8bb822b949a03b645d2be412d3da9f9ded328311c632252c16f" exitCode=0 Oct 03 15:36:29 crc kubenswrapper[4959]: I1003 15:36:29.521330 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" event={"ID":"7b050197-797a-4c1d-8e40-530801ae7bd3","Type":"ContainerDied","Data":"c0810cb4af6de8bb822b949a03b645d2be412d3da9f9ded328311c632252c16f"} Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.091815 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.243173 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsslj\" (UniqueName: \"kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj\") pod \"7b050197-797a-4c1d-8e40-530801ae7bd3\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.243590 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory\") pod \"7b050197-797a-4c1d-8e40-530801ae7bd3\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.243693 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph\") pod \"7b050197-797a-4c1d-8e40-530801ae7bd3\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.243786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key\") pod \"7b050197-797a-4c1d-8e40-530801ae7bd3\" (UID: \"7b050197-797a-4c1d-8e40-530801ae7bd3\") " Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.248741 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj" (OuterVolumeSpecName: "kube-api-access-lsslj") pod "7b050197-797a-4c1d-8e40-530801ae7bd3" (UID: "7b050197-797a-4c1d-8e40-530801ae7bd3"). InnerVolumeSpecName "kube-api-access-lsslj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.257405 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph" (OuterVolumeSpecName: "ceph") pod "7b050197-797a-4c1d-8e40-530801ae7bd3" (UID: "7b050197-797a-4c1d-8e40-530801ae7bd3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.280635 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory" (OuterVolumeSpecName: "inventory") pod "7b050197-797a-4c1d-8e40-530801ae7bd3" (UID: "7b050197-797a-4c1d-8e40-530801ae7bd3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.287128 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b050197-797a-4c1d-8e40-530801ae7bd3" (UID: "7b050197-797a-4c1d-8e40-530801ae7bd3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.346544 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.346610 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsslj\" (UniqueName: \"kubernetes.io/projected/7b050197-797a-4c1d-8e40-530801ae7bd3-kube-api-access-lsslj\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.346635 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.346654 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b050197-797a-4c1d-8e40-530801ae7bd3-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.548665 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" event={"ID":"7b050197-797a-4c1d-8e40-530801ae7bd3","Type":"ContainerDied","Data":"e22893defb781f84374c141507db6dde65d7638f44bba3ef7e32b844732078ff"} Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.548724 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e22893defb781f84374c141507db6dde65d7638f44bba3ef7e32b844732078ff" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.548797 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-jlh5w" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.639468 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vm92f"] Oct 03 15:36:31 crc kubenswrapper[4959]: E1003 15:36:31.640066 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="registry-server" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640112 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="registry-server" Oct 03 15:36:31 crc kubenswrapper[4959]: E1003 15:36:31.640152 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="extract-content" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640161 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="extract-content" Oct 03 15:36:31 crc kubenswrapper[4959]: E1003 15:36:31.640183 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="extract-utilities" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640209 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="extract-utilities" Oct 03 15:36:31 crc kubenswrapper[4959]: E1003 15:36:31.640227 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b050197-797a-4c1d-8e40-530801ae7bd3" containerName="download-cache-openstack-openstack-cell1" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640235 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b050197-797a-4c1d-8e40-530801ae7bd3" containerName="download-cache-openstack-openstack-cell1" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640478 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d111ecf-fdc6-41e7-871c-fcb6ddaddf5e" containerName="registry-server" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.640507 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b050197-797a-4c1d-8e40-530801ae7bd3" containerName="download-cache-openstack-openstack-cell1" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.641526 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.645440 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.645458 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.645659 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.645910 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.648760 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vm92f"] Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.755410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.755529 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.755748 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.755808 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq26q\" (UniqueName: \"kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.865266 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.865333 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq26q\" (UniqueName: \"kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.865642 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.865892 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.867552 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.879878 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.880881 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.882385 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.890838 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.895884 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq26q\" (UniqueName: \"kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q\") pod \"configure-network-openstack-openstack-cell1-vm92f\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.969355 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:36:31 crc kubenswrapper[4959]: I1003 15:36:31.978037 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:36:33 crc kubenswrapper[4959]: I1003 15:36:32.624063 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vm92f"] Oct 03 15:36:33 crc kubenswrapper[4959]: I1003 15:36:32.802032 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:36:33 crc kubenswrapper[4959]: I1003 15:36:33.571129 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" event={"ID":"8d721f05-6907-49fa-87ed-de92d1fa8139","Type":"ContainerStarted","Data":"6d35a7976ebfe57d5ca56bed094d885c1c8eb0a107f5ece88af83583e66814b2"} Oct 03 15:36:33 crc kubenswrapper[4959]: I1003 15:36:33.571522 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" event={"ID":"8d721f05-6907-49fa-87ed-de92d1fa8139","Type":"ContainerStarted","Data":"81b6a4dede4fc0d363fee0eb5017ec19e5ed74ca12263920d601a0be09921a3a"} Oct 03 15:36:33 crc kubenswrapper[4959]: I1003 15:36:33.603751 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" podStartSLOduration=2.435684421 podStartE2EDuration="2.603724818s" podCreationTimestamp="2025-10-03 15:36:31 +0000 UTC" firstStartedPulling="2025-10-03 15:36:32.631134603 +0000 UTC m=+7561.834478020" lastFinishedPulling="2025-10-03 15:36:32.79917499 +0000 UTC m=+7562.002518417" observedRunningTime="2025-10-03 15:36:33.591774025 +0000 UTC m=+7562.795117442" watchObservedRunningTime="2025-10-03 15:36:33.603724818 +0000 UTC m=+7562.807068275" Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.044296 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.044637 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.044684 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.045568 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.045921 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030" gracePeriod=600 Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.613396 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030" exitCode=0 Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.613508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030"} Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.613798 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4"} Oct 03 15:36:36 crc kubenswrapper[4959]: I1003 15:36:36.613829 4959 scope.go:117] "RemoveContainer" containerID="bd332d84898f219aeed8bc3654d845e7e990d74a38acf29d3c45dd04e7704431" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.279011 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.282470 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.299382 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.408406 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.408480 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577nt\" (UniqueName: \"kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.408579 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.510976 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.511367 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577nt\" (UniqueName: \"kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.511579 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.511597 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.511988 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.535306 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577nt\" (UniqueName: \"kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt\") pod \"redhat-marketplace-snndh\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:53 crc kubenswrapper[4959]: I1003 15:37:53.601999 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.079135 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.594313 4959 generic.go:334] "Generic (PLEG): container finished" podID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerID="76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920" exitCode=0 Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.594353 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerDied","Data":"76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920"} Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.594612 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerStarted","Data":"e18a29c6824a62af4c8a4f1aa546664548e7ce933a4b31ee7394b78a0818ecb1"} Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.597296 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.597605 4959 generic.go:334] "Generic (PLEG): container finished" podID="8d721f05-6907-49fa-87ed-de92d1fa8139" containerID="6d35a7976ebfe57d5ca56bed094d885c1c8eb0a107f5ece88af83583e66814b2" exitCode=0 Oct 03 15:37:54 crc kubenswrapper[4959]: I1003 15:37:54.597630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" event={"ID":"8d721f05-6907-49fa-87ed-de92d1fa8139","Type":"ContainerDied","Data":"6d35a7976ebfe57d5ca56bed094d885c1c8eb0a107f5ece88af83583e66814b2"} Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.088474 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.172261 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key\") pod \"8d721f05-6907-49fa-87ed-de92d1fa8139\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.172331 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph\") pod \"8d721f05-6907-49fa-87ed-de92d1fa8139\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.172393 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq26q\" (UniqueName: \"kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q\") pod \"8d721f05-6907-49fa-87ed-de92d1fa8139\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.172513 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory\") pod \"8d721f05-6907-49fa-87ed-de92d1fa8139\" (UID: \"8d721f05-6907-49fa-87ed-de92d1fa8139\") " Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.184581 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph" (OuterVolumeSpecName: "ceph") pod "8d721f05-6907-49fa-87ed-de92d1fa8139" (UID: "8d721f05-6907-49fa-87ed-de92d1fa8139"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.184663 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q" (OuterVolumeSpecName: "kube-api-access-vq26q") pod "8d721f05-6907-49fa-87ed-de92d1fa8139" (UID: "8d721f05-6907-49fa-87ed-de92d1fa8139"). InnerVolumeSpecName "kube-api-access-vq26q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.206160 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory" (OuterVolumeSpecName: "inventory") pod "8d721f05-6907-49fa-87ed-de92d1fa8139" (UID: "8d721f05-6907-49fa-87ed-de92d1fa8139"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.207619 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8d721f05-6907-49fa-87ed-de92d1fa8139" (UID: "8d721f05-6907-49fa-87ed-de92d1fa8139"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.274983 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq26q\" (UniqueName: \"kubernetes.io/projected/8d721f05-6907-49fa-87ed-de92d1fa8139-kube-api-access-vq26q\") on node \"crc\" DevicePath \"\"" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.275022 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.275038 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.275052 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8d721f05-6907-49fa-87ed-de92d1fa8139-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.620538 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" event={"ID":"8d721f05-6907-49fa-87ed-de92d1fa8139","Type":"ContainerDied","Data":"81b6a4dede4fc0d363fee0eb5017ec19e5ed74ca12263920d601a0be09921a3a"} Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.620588 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81b6a4dede4fc0d363fee0eb5017ec19e5ed74ca12263920d601a0be09921a3a" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.620659 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vm92f" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.623836 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerStarted","Data":"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d"} Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.750853 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-kw7hh"] Oct 03 15:37:56 crc kubenswrapper[4959]: E1003 15:37:56.751441 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d721f05-6907-49fa-87ed-de92d1fa8139" containerName="configure-network-openstack-openstack-cell1" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.751466 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d721f05-6907-49fa-87ed-de92d1fa8139" containerName="configure-network-openstack-openstack-cell1" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.751756 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d721f05-6907-49fa-87ed-de92d1fa8139" containerName="configure-network-openstack-openstack-cell1" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.752622 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.755209 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.756570 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.756699 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.756932 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.774078 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-kw7hh"] Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.792306 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.792346 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.792394 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ql9n\" (UniqueName: \"kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.793892 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: E1003 15:37:56.865192 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d721f05_6907_49fa_87ed_de92d1fa8139.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d721f05_6907_49fa_87ed_de92d1fa8139.slice/crio-81b6a4dede4fc0d363fee0eb5017ec19e5ed74ca12263920d601a0be09921a3a\": RecentStats: unable to find data in memory cache]" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.895763 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.896292 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.896314 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.896348 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ql9n\" (UniqueName: \"kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.900958 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.911580 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.912075 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:56 crc kubenswrapper[4959]: I1003 15:37:56.916796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ql9n\" (UniqueName: \"kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n\") pod \"validate-network-openstack-openstack-cell1-kw7hh\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:57 crc kubenswrapper[4959]: I1003 15:37:57.097662 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:37:57 crc kubenswrapper[4959]: I1003 15:37:57.644953 4959 generic.go:334] "Generic (PLEG): container finished" podID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerID="4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d" exitCode=0 Oct 03 15:37:57 crc kubenswrapper[4959]: I1003 15:37:57.645006 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerDied","Data":"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d"} Oct 03 15:37:57 crc kubenswrapper[4959]: I1003 15:37:57.744608 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-kw7hh"] Oct 03 15:37:58 crc kubenswrapper[4959]: I1003 15:37:58.666222 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" event={"ID":"33563547-3b0f-4647-8aff-1a068907aa2d","Type":"ContainerStarted","Data":"fe081f65607c0e88d61a7cea6934756a30cc01d795e9b873cae5172c663d61c3"} Oct 03 15:37:59 crc kubenswrapper[4959]: I1003 15:37:59.699593 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerStarted","Data":"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e"} Oct 03 15:37:59 crc kubenswrapper[4959]: I1003 15:37:59.700790 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" event={"ID":"33563547-3b0f-4647-8aff-1a068907aa2d","Type":"ContainerStarted","Data":"c89d54341c9e735aa2cc6aebdb49d73d147dd62f1d1871f261438a3b48ef7521"} Oct 03 15:37:59 crc kubenswrapper[4959]: I1003 15:37:59.711310 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-snndh" podStartSLOduration=2.8058665449999998 podStartE2EDuration="6.711286892s" podCreationTimestamp="2025-10-03 15:37:53 +0000 UTC" firstStartedPulling="2025-10-03 15:37:54.597030026 +0000 UTC m=+7643.800373443" lastFinishedPulling="2025-10-03 15:37:58.502450373 +0000 UTC m=+7647.705793790" observedRunningTime="2025-10-03 15:37:59.70549401 +0000 UTC m=+7648.908837427" watchObservedRunningTime="2025-10-03 15:37:59.711286892 +0000 UTC m=+7648.914630319" Oct 03 15:37:59 crc kubenswrapper[4959]: I1003 15:37:59.747812 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" podStartSLOduration=3.061369945 podStartE2EDuration="3.747787154s" podCreationTimestamp="2025-10-03 15:37:56 +0000 UTC" firstStartedPulling="2025-10-03 15:37:57.745336145 +0000 UTC m=+7646.948679562" lastFinishedPulling="2025-10-03 15:37:58.431753334 +0000 UTC m=+7647.635096771" observedRunningTime="2025-10-03 15:37:59.733488265 +0000 UTC m=+7648.936831692" watchObservedRunningTime="2025-10-03 15:37:59.747787154 +0000 UTC m=+7648.951130591" Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.603174 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.603695 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.706867 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.741959 4959 generic.go:334] "Generic (PLEG): container finished" podID="33563547-3b0f-4647-8aff-1a068907aa2d" containerID="c89d54341c9e735aa2cc6aebdb49d73d147dd62f1d1871f261438a3b48ef7521" exitCode=0 Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.742034 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" event={"ID":"33563547-3b0f-4647-8aff-1a068907aa2d","Type":"ContainerDied","Data":"c89d54341c9e735aa2cc6aebdb49d73d147dd62f1d1871f261438a3b48ef7521"} Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.788438 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:03 crc kubenswrapper[4959]: I1003 15:38:03.944204 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.273261 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.391854 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key\") pod \"33563547-3b0f-4647-8aff-1a068907aa2d\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.392148 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ql9n\" (UniqueName: \"kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n\") pod \"33563547-3b0f-4647-8aff-1a068907aa2d\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.392381 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory\") pod \"33563547-3b0f-4647-8aff-1a068907aa2d\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.392494 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph\") pod \"33563547-3b0f-4647-8aff-1a068907aa2d\" (UID: \"33563547-3b0f-4647-8aff-1a068907aa2d\") " Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.397564 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n" (OuterVolumeSpecName: "kube-api-access-4ql9n") pod "33563547-3b0f-4647-8aff-1a068907aa2d" (UID: "33563547-3b0f-4647-8aff-1a068907aa2d"). InnerVolumeSpecName "kube-api-access-4ql9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.412254 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph" (OuterVolumeSpecName: "ceph") pod "33563547-3b0f-4647-8aff-1a068907aa2d" (UID: "33563547-3b0f-4647-8aff-1a068907aa2d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.423686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory" (OuterVolumeSpecName: "inventory") pod "33563547-3b0f-4647-8aff-1a068907aa2d" (UID: "33563547-3b0f-4647-8aff-1a068907aa2d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.424053 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "33563547-3b0f-4647-8aff-1a068907aa2d" (UID: "33563547-3b0f-4647-8aff-1a068907aa2d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.496283 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ql9n\" (UniqueName: \"kubernetes.io/projected/33563547-3b0f-4647-8aff-1a068907aa2d-kube-api-access-4ql9n\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.496311 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.496326 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.496338 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33563547-3b0f-4647-8aff-1a068907aa2d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.767405 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" event={"ID":"33563547-3b0f-4647-8aff-1a068907aa2d","Type":"ContainerDied","Data":"fe081f65607c0e88d61a7cea6934756a30cc01d795e9b873cae5172c663d61c3"} Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.767768 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe081f65607c0e88d61a7cea6934756a30cc01d795e9b873cae5172c663d61c3" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.767510 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-snndh" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="registry-server" containerID="cri-o://ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e" gracePeriod=2 Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.767467 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-kw7hh" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.849610 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-b85rz"] Oct 03 15:38:05 crc kubenswrapper[4959]: E1003 15:38:05.850003 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33563547-3b0f-4647-8aff-1a068907aa2d" containerName="validate-network-openstack-openstack-cell1" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.850021 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="33563547-3b0f-4647-8aff-1a068907aa2d" containerName="validate-network-openstack-openstack-cell1" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.850367 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="33563547-3b0f-4647-8aff-1a068907aa2d" containerName="validate-network-openstack-openstack-cell1" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.851059 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.853415 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.853515 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.854639 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.855623 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:38:05 crc kubenswrapper[4959]: I1003 15:38:05.868588 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-b85rz"] Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.019160 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.019488 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.019563 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l74mb\" (UniqueName: \"kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.019629 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.121760 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.121799 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.121865 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l74mb\" (UniqueName: \"kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.121921 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.126001 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.127662 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.128275 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.139207 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l74mb\" (UniqueName: \"kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb\") pod \"install-os-openstack-openstack-cell1-b85rz\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.255463 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.366064 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.529104 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577nt\" (UniqueName: \"kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt\") pod \"4c90f140-aa07-4045-a2ad-867133d9bb92\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.529206 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities\") pod \"4c90f140-aa07-4045-a2ad-867133d9bb92\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.529407 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content\") pod \"4c90f140-aa07-4045-a2ad-867133d9bb92\" (UID: \"4c90f140-aa07-4045-a2ad-867133d9bb92\") " Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.530387 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities" (OuterVolumeSpecName: "utilities") pod "4c90f140-aa07-4045-a2ad-867133d9bb92" (UID: "4c90f140-aa07-4045-a2ad-867133d9bb92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.534984 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt" (OuterVolumeSpecName: "kube-api-access-577nt") pod "4c90f140-aa07-4045-a2ad-867133d9bb92" (UID: "4c90f140-aa07-4045-a2ad-867133d9bb92"). InnerVolumeSpecName "kube-api-access-577nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.545076 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c90f140-aa07-4045-a2ad-867133d9bb92" (UID: "4c90f140-aa07-4045-a2ad-867133d9bb92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.631636 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.631669 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577nt\" (UniqueName: \"kubernetes.io/projected/4c90f140-aa07-4045-a2ad-867133d9bb92-kube-api-access-577nt\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.631699 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c90f140-aa07-4045-a2ad-867133d9bb92-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.778364 4959 generic.go:334] "Generic (PLEG): container finished" podID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerID="ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e" exitCode=0 Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.778407 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerDied","Data":"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e"} Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.778436 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-snndh" event={"ID":"4c90f140-aa07-4045-a2ad-867133d9bb92","Type":"ContainerDied","Data":"e18a29c6824a62af4c8a4f1aa546664548e7ce933a4b31ee7394b78a0818ecb1"} Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.778453 4959 scope.go:117] "RemoveContainer" containerID="ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.778584 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-snndh" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.809641 4959 scope.go:117] "RemoveContainer" containerID="4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.824788 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.843046 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-snndh"] Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.850185 4959 scope.go:117] "RemoveContainer" containerID="76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.851323 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-b85rz"] Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.872499 4959 scope.go:117] "RemoveContainer" containerID="ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e" Oct 03 15:38:06 crc kubenswrapper[4959]: E1003 15:38:06.873045 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e\": container with ID starting with ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e not found: ID does not exist" containerID="ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.873077 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e"} err="failed to get container status \"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e\": rpc error: code = NotFound desc = could not find container \"ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e\": container with ID starting with ae91319c923134685ae3a4a10876b810045f268d86dc65d752c609a73b28c27e not found: ID does not exist" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.873098 4959 scope.go:117] "RemoveContainer" containerID="4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d" Oct 03 15:38:06 crc kubenswrapper[4959]: E1003 15:38:06.873556 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d\": container with ID starting with 4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d not found: ID does not exist" containerID="4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.873580 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d"} err="failed to get container status \"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d\": rpc error: code = NotFound desc = could not find container \"4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d\": container with ID starting with 4a1d7f63f175a856c37d72c38a7205a868bc3e3ab55b2e3e2dc0cdb85541f97d not found: ID does not exist" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.873593 4959 scope.go:117] "RemoveContainer" containerID="76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920" Oct 03 15:38:06 crc kubenswrapper[4959]: E1003 15:38:06.873797 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920\": container with ID starting with 76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920 not found: ID does not exist" containerID="76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920" Oct 03 15:38:06 crc kubenswrapper[4959]: I1003 15:38:06.873814 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920"} err="failed to get container status \"76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920\": rpc error: code = NotFound desc = could not find container \"76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920\": container with ID starting with 76067b36bcd3a645c989e28e57db6ddc955d407230e787e5226e5fa48a3fe920 not found: ID does not exist" Oct 03 15:38:07 crc kubenswrapper[4959]: I1003 15:38:07.697075 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" path="/var/lib/kubelet/pods/4c90f140-aa07-4045-a2ad-867133d9bb92/volumes" Oct 03 15:38:07 crc kubenswrapper[4959]: I1003 15:38:07.789097 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-b85rz" event={"ID":"f3102942-6ad8-45bf-a609-90bdd43c2176","Type":"ContainerStarted","Data":"5f34d27bb6f1af411d37e9d0dcc690de67b33d88d413425ae35cbad9b03831e5"} Oct 03 15:38:07 crc kubenswrapper[4959]: I1003 15:38:07.789141 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-b85rz" event={"ID":"f3102942-6ad8-45bf-a609-90bdd43c2176","Type":"ContainerStarted","Data":"39679f936a02a32a86232675b7538f57f42d66985c6282b098dcfef03eb19c7b"} Oct 03 15:38:07 crc kubenswrapper[4959]: I1003 15:38:07.803655 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-b85rz" podStartSLOduration=2.631258511 podStartE2EDuration="2.803638695s" podCreationTimestamp="2025-10-03 15:38:05 +0000 UTC" firstStartedPulling="2025-10-03 15:38:06.860517571 +0000 UTC m=+7656.063860988" lastFinishedPulling="2025-10-03 15:38:07.032897755 +0000 UTC m=+7656.236241172" observedRunningTime="2025-10-03 15:38:07.801024761 +0000 UTC m=+7657.004368178" watchObservedRunningTime="2025-10-03 15:38:07.803638695 +0000 UTC m=+7657.006982112" Oct 03 15:38:36 crc kubenswrapper[4959]: I1003 15:38:36.044259 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:38:36 crc kubenswrapper[4959]: I1003 15:38:36.044925 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:38:53 crc kubenswrapper[4959]: I1003 15:38:53.380777 4959 generic.go:334] "Generic (PLEG): container finished" podID="f3102942-6ad8-45bf-a609-90bdd43c2176" containerID="5f34d27bb6f1af411d37e9d0dcc690de67b33d88d413425ae35cbad9b03831e5" exitCode=0 Oct 03 15:38:53 crc kubenswrapper[4959]: I1003 15:38:53.380885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-b85rz" event={"ID":"f3102942-6ad8-45bf-a609-90bdd43c2176","Type":"ContainerDied","Data":"5f34d27bb6f1af411d37e9d0dcc690de67b33d88d413425ae35cbad9b03831e5"} Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.862018 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.988960 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key\") pod \"f3102942-6ad8-45bf-a609-90bdd43c2176\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.989102 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory\") pod \"f3102942-6ad8-45bf-a609-90bdd43c2176\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.989142 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l74mb\" (UniqueName: \"kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb\") pod \"f3102942-6ad8-45bf-a609-90bdd43c2176\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.989943 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph\") pod \"f3102942-6ad8-45bf-a609-90bdd43c2176\" (UID: \"f3102942-6ad8-45bf-a609-90bdd43c2176\") " Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.996086 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph" (OuterVolumeSpecName: "ceph") pod "f3102942-6ad8-45bf-a609-90bdd43c2176" (UID: "f3102942-6ad8-45bf-a609-90bdd43c2176"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:54 crc kubenswrapper[4959]: I1003 15:38:54.996230 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb" (OuterVolumeSpecName: "kube-api-access-l74mb") pod "f3102942-6ad8-45bf-a609-90bdd43c2176" (UID: "f3102942-6ad8-45bf-a609-90bdd43c2176"). InnerVolumeSpecName "kube-api-access-l74mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.034821 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f3102942-6ad8-45bf-a609-90bdd43c2176" (UID: "f3102942-6ad8-45bf-a609-90bdd43c2176"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.048516 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory" (OuterVolumeSpecName: "inventory") pod "f3102942-6ad8-45bf-a609-90bdd43c2176" (UID: "f3102942-6ad8-45bf-a609-90bdd43c2176"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.092141 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.092174 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.092185 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l74mb\" (UniqueName: \"kubernetes.io/projected/f3102942-6ad8-45bf-a609-90bdd43c2176-kube-api-access-l74mb\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.092212 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3102942-6ad8-45bf-a609-90bdd43c2176-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.416130 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-b85rz" event={"ID":"f3102942-6ad8-45bf-a609-90bdd43c2176","Type":"ContainerDied","Data":"39679f936a02a32a86232675b7538f57f42d66985c6282b098dcfef03eb19c7b"} Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.416462 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39679f936a02a32a86232675b7538f57f42d66985c6282b098dcfef03eb19c7b" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.416283 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-b85rz" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.508251 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hnc2k"] Oct 03 15:38:55 crc kubenswrapper[4959]: E1003 15:38:55.511267 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="extract-content" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511297 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="extract-content" Oct 03 15:38:55 crc kubenswrapper[4959]: E1003 15:38:55.511323 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3102942-6ad8-45bf-a609-90bdd43c2176" containerName="install-os-openstack-openstack-cell1" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511329 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3102942-6ad8-45bf-a609-90bdd43c2176" containerName="install-os-openstack-openstack-cell1" Oct 03 15:38:55 crc kubenswrapper[4959]: E1003 15:38:55.511346 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="registry-server" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511351 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="registry-server" Oct 03 15:38:55 crc kubenswrapper[4959]: E1003 15:38:55.511391 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="extract-utilities" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511396 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="extract-utilities" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511582 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c90f140-aa07-4045-a2ad-867133d9bb92" containerName="registry-server" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.511611 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3102942-6ad8-45bf-a609-90bdd43c2176" containerName="install-os-openstack-openstack-cell1" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.512417 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.516119 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.516654 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.516713 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.518558 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.522546 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hnc2k"] Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.602023 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.602070 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7ckm\" (UniqueName: \"kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.602149 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.602316 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.705313 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.705475 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7ckm\" (UniqueName: \"kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.705702 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.706011 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.709857 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.711117 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.714728 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.735825 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7ckm\" (UniqueName: \"kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm\") pod \"configure-os-openstack-openstack-cell1-hnc2k\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:55 crc kubenswrapper[4959]: I1003 15:38:55.831586 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:38:56 crc kubenswrapper[4959]: I1003 15:38:56.477928 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-hnc2k"] Oct 03 15:38:57 crc kubenswrapper[4959]: I1003 15:38:57.451900 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" event={"ID":"b3b15986-a162-4643-a76e-67d33d624642","Type":"ContainerStarted","Data":"c1e09f4e08b9e655ea480134d33b071e4c4fdb4d6a4eb8bb34e0b770b9368620"} Oct 03 15:38:57 crc kubenswrapper[4959]: I1003 15:38:57.452479 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" event={"ID":"b3b15986-a162-4643-a76e-67d33d624642","Type":"ContainerStarted","Data":"0a3f7ca3185c6a336ee5199e75e34afacd97d9487f9b7a0c7f520808af09aa49"} Oct 03 15:38:57 crc kubenswrapper[4959]: I1003 15:38:57.487958 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" podStartSLOduration=2.309567482 podStartE2EDuration="2.487942153s" podCreationTimestamp="2025-10-03 15:38:55 +0000 UTC" firstStartedPulling="2025-10-03 15:38:56.466118375 +0000 UTC m=+7705.669461792" lastFinishedPulling="2025-10-03 15:38:56.644493046 +0000 UTC m=+7705.847836463" observedRunningTime="2025-10-03 15:38:57.484304954 +0000 UTC m=+7706.687648371" watchObservedRunningTime="2025-10-03 15:38:57.487942153 +0000 UTC m=+7706.691285580" Oct 03 15:39:06 crc kubenswrapper[4959]: I1003 15:39:06.044809 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:39:06 crc kubenswrapper[4959]: I1003 15:39:06.045586 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.044013 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.046297 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.046498 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.048067 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.048370 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" gracePeriod=600 Oct 03 15:39:36 crc kubenswrapper[4959]: E1003 15:39:36.170053 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.927084 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" exitCode=0 Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.927135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4"} Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.927515 4959 scope.go:117] "RemoveContainer" containerID="11aa88aecf2a3928c4475e65ccb3b7fad7192ec231b4f71b723d4f9586b7b030" Oct 03 15:39:36 crc kubenswrapper[4959]: I1003 15:39:36.927910 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:39:36 crc kubenswrapper[4959]: E1003 15:39:36.928205 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:39:43 crc kubenswrapper[4959]: I1003 15:39:43.007020 4959 generic.go:334] "Generic (PLEG): container finished" podID="b3b15986-a162-4643-a76e-67d33d624642" containerID="c1e09f4e08b9e655ea480134d33b071e4c4fdb4d6a4eb8bb34e0b770b9368620" exitCode=0 Oct 03 15:39:43 crc kubenswrapper[4959]: I1003 15:39:43.007112 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" event={"ID":"b3b15986-a162-4643-a76e-67d33d624642","Type":"ContainerDied","Data":"c1e09f4e08b9e655ea480134d33b071e4c4fdb4d6a4eb8bb34e0b770b9368620"} Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.554623 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.682029 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7ckm\" (UniqueName: \"kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm\") pod \"b3b15986-a162-4643-a76e-67d33d624642\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.682069 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph\") pod \"b3b15986-a162-4643-a76e-67d33d624642\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.682147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory\") pod \"b3b15986-a162-4643-a76e-67d33d624642\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.682314 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key\") pod \"b3b15986-a162-4643-a76e-67d33d624642\" (UID: \"b3b15986-a162-4643-a76e-67d33d624642\") " Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.695569 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm" (OuterVolumeSpecName: "kube-api-access-m7ckm") pod "b3b15986-a162-4643-a76e-67d33d624642" (UID: "b3b15986-a162-4643-a76e-67d33d624642"). InnerVolumeSpecName "kube-api-access-m7ckm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.695619 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph" (OuterVolumeSpecName: "ceph") pod "b3b15986-a162-4643-a76e-67d33d624642" (UID: "b3b15986-a162-4643-a76e-67d33d624642"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.718335 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3b15986-a162-4643-a76e-67d33d624642" (UID: "b3b15986-a162-4643-a76e-67d33d624642"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.741295 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory" (OuterVolumeSpecName: "inventory") pod "b3b15986-a162-4643-a76e-67d33d624642" (UID: "b3b15986-a162-4643-a76e-67d33d624642"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.784844 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7ckm\" (UniqueName: \"kubernetes.io/projected/b3b15986-a162-4643-a76e-67d33d624642-kube-api-access-m7ckm\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.784900 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.784923 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:44 crc kubenswrapper[4959]: I1003 15:39:44.784942 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3b15986-a162-4643-a76e-67d33d624642-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.030682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" event={"ID":"b3b15986-a162-4643-a76e-67d33d624642","Type":"ContainerDied","Data":"0a3f7ca3185c6a336ee5199e75e34afacd97d9487f9b7a0c7f520808af09aa49"} Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.031137 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a3f7ca3185c6a336ee5199e75e34afacd97d9487f9b7a0c7f520808af09aa49" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.030849 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-hnc2k" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.136222 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-j9ckb"] Oct 03 15:39:45 crc kubenswrapper[4959]: E1003 15:39:45.136759 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b15986-a162-4643-a76e-67d33d624642" containerName="configure-os-openstack-openstack-cell1" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.136780 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b15986-a162-4643-a76e-67d33d624642" containerName="configure-os-openstack-openstack-cell1" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.137044 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b15986-a162-4643-a76e-67d33d624642" containerName="configure-os-openstack-openstack-cell1" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.137947 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.140719 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.140944 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.141085 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.141179 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.150496 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-j9ckb"] Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.297362 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.297681 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.297813 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshsg\" (UniqueName: \"kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.297893 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.400327 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.400472 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.400555 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshsg\" (UniqueName: \"kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.400594 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.405880 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.407811 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.408935 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.430881 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshsg\" (UniqueName: \"kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg\") pod \"ssh-known-hosts-openstack-j9ckb\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:45 crc kubenswrapper[4959]: I1003 15:39:45.457799 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:46 crc kubenswrapper[4959]: I1003 15:39:46.014310 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-j9ckb"] Oct 03 15:39:46 crc kubenswrapper[4959]: I1003 15:39:46.047080 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-j9ckb" event={"ID":"521fed84-00a8-449a-a036-dda24250ddc4","Type":"ContainerStarted","Data":"821830aa0261132facbd133742ce2cdbd7ed176ff4b599a53817733f89d921df"} Oct 03 15:39:47 crc kubenswrapper[4959]: I1003 15:39:47.058036 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-j9ckb" event={"ID":"521fed84-00a8-449a-a036-dda24250ddc4","Type":"ContainerStarted","Data":"3a26ec761df10c51e8fac2e24a8e8b403c43aceca3fcf27d611caa6763992225"} Oct 03 15:39:47 crc kubenswrapper[4959]: I1003 15:39:47.090726 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-j9ckb" podStartSLOduration=1.934456999 podStartE2EDuration="2.090701678s" podCreationTimestamp="2025-10-03 15:39:45 +0000 UTC" firstStartedPulling="2025-10-03 15:39:46.023309586 +0000 UTC m=+7755.226653003" lastFinishedPulling="2025-10-03 15:39:46.179554255 +0000 UTC m=+7755.382897682" observedRunningTime="2025-10-03 15:39:47.086041264 +0000 UTC m=+7756.289384681" watchObservedRunningTime="2025-10-03 15:39:47.090701678 +0000 UTC m=+7756.294045095" Oct 03 15:39:49 crc kubenswrapper[4959]: I1003 15:39:49.686843 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:39:49 crc kubenswrapper[4959]: E1003 15:39:49.688353 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:39:56 crc kubenswrapper[4959]: I1003 15:39:56.186323 4959 generic.go:334] "Generic (PLEG): container finished" podID="521fed84-00a8-449a-a036-dda24250ddc4" containerID="3a26ec761df10c51e8fac2e24a8e8b403c43aceca3fcf27d611caa6763992225" exitCode=0 Oct 03 15:39:56 crc kubenswrapper[4959]: I1003 15:39:56.186444 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-j9ckb" event={"ID":"521fed84-00a8-449a-a036-dda24250ddc4","Type":"ContainerDied","Data":"3a26ec761df10c51e8fac2e24a8e8b403c43aceca3fcf27d611caa6763992225"} Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.746722 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.821865 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph\") pod \"521fed84-00a8-449a-a036-dda24250ddc4\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.821922 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0\") pod \"521fed84-00a8-449a-a036-dda24250ddc4\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.822248 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pshsg\" (UniqueName: \"kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg\") pod \"521fed84-00a8-449a-a036-dda24250ddc4\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.822380 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1\") pod \"521fed84-00a8-449a-a036-dda24250ddc4\" (UID: \"521fed84-00a8-449a-a036-dda24250ddc4\") " Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.829593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg" (OuterVolumeSpecName: "kube-api-access-pshsg") pod "521fed84-00a8-449a-a036-dda24250ddc4" (UID: "521fed84-00a8-449a-a036-dda24250ddc4"). InnerVolumeSpecName "kube-api-access-pshsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.833751 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph" (OuterVolumeSpecName: "ceph") pod "521fed84-00a8-449a-a036-dda24250ddc4" (UID: "521fed84-00a8-449a-a036-dda24250ddc4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.852419 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "521fed84-00a8-449a-a036-dda24250ddc4" (UID: "521fed84-00a8-449a-a036-dda24250ddc4"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.859120 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "521fed84-00a8-449a-a036-dda24250ddc4" (UID: "521fed84-00a8-449a-a036-dda24250ddc4"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.925745 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.925780 4959 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.925794 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pshsg\" (UniqueName: \"kubernetes.io/projected/521fed84-00a8-449a-a036-dda24250ddc4-kube-api-access-pshsg\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:57 crc kubenswrapper[4959]: I1003 15:39:57.925802 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/521fed84-00a8-449a-a036-dda24250ddc4-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.216960 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-j9ckb" event={"ID":"521fed84-00a8-449a-a036-dda24250ddc4","Type":"ContainerDied","Data":"821830aa0261132facbd133742ce2cdbd7ed176ff4b599a53817733f89d921df"} Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.217000 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="821830aa0261132facbd133742ce2cdbd7ed176ff4b599a53817733f89d921df" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.217024 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-j9ckb" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.336274 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-c9lbc"] Oct 03 15:39:58 crc kubenswrapper[4959]: E1003 15:39:58.337457 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521fed84-00a8-449a-a036-dda24250ddc4" containerName="ssh-known-hosts-openstack" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.337500 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="521fed84-00a8-449a-a036-dda24250ddc4" containerName="ssh-known-hosts-openstack" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.338327 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="521fed84-00a8-449a-a036-dda24250ddc4" containerName="ssh-known-hosts-openstack" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.340048 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.351317 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-c9lbc"] Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.362873 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.362621 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.363365 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.363736 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.443133 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxbp6\" (UniqueName: \"kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.443293 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.443340 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.443369 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.546377 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxbp6\" (UniqueName: \"kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.546456 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.546481 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.546496 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.551810 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.552726 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.554451 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.565292 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxbp6\" (UniqueName: \"kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6\") pod \"run-os-openstack-openstack-cell1-c9lbc\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:58 crc kubenswrapper[4959]: I1003 15:39:58.688662 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:39:59 crc kubenswrapper[4959]: I1003 15:39:59.249804 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-c9lbc"] Oct 03 15:39:59 crc kubenswrapper[4959]: W1003 15:39:59.259733 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf37d30c9_da2b_4894_9f1f_30c37f9b0b19.slice/crio-df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab WatchSource:0}: Error finding container df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab: Status 404 returned error can't find the container with id df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab Oct 03 15:40:00 crc kubenswrapper[4959]: I1003 15:40:00.244477 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" event={"ID":"f37d30c9-da2b-4894-9f1f-30c37f9b0b19","Type":"ContainerStarted","Data":"9e149cade3f028054296deabb7e2a6080f33bf275ef52c7ea46307b86a9087cf"} Oct 03 15:40:00 crc kubenswrapper[4959]: I1003 15:40:00.244791 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" event={"ID":"f37d30c9-da2b-4894-9f1f-30c37f9b0b19","Type":"ContainerStarted","Data":"df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab"} Oct 03 15:40:00 crc kubenswrapper[4959]: I1003 15:40:00.277702 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" podStartSLOduration=2.099138361 podStartE2EDuration="2.277678716s" podCreationTimestamp="2025-10-03 15:39:58 +0000 UTC" firstStartedPulling="2025-10-03 15:39:59.263635429 +0000 UTC m=+7768.466978856" lastFinishedPulling="2025-10-03 15:39:59.442175794 +0000 UTC m=+7768.645519211" observedRunningTime="2025-10-03 15:40:00.263956382 +0000 UTC m=+7769.467299799" watchObservedRunningTime="2025-10-03 15:40:00.277678716 +0000 UTC m=+7769.481022133" Oct 03 15:40:04 crc kubenswrapper[4959]: I1003 15:40:04.686525 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:40:04 crc kubenswrapper[4959]: E1003 15:40:04.687283 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:40:08 crc kubenswrapper[4959]: I1003 15:40:08.360116 4959 generic.go:334] "Generic (PLEG): container finished" podID="f37d30c9-da2b-4894-9f1f-30c37f9b0b19" containerID="9e149cade3f028054296deabb7e2a6080f33bf275ef52c7ea46307b86a9087cf" exitCode=0 Oct 03 15:40:08 crc kubenswrapper[4959]: I1003 15:40:08.360275 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" event={"ID":"f37d30c9-da2b-4894-9f1f-30c37f9b0b19","Type":"ContainerDied","Data":"9e149cade3f028054296deabb7e2a6080f33bf275ef52c7ea46307b86a9087cf"} Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.823353 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.912062 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph\") pod \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.912707 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory\") pod \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.912885 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxbp6\" (UniqueName: \"kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6\") pod \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.913022 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key\") pod \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\" (UID: \"f37d30c9-da2b-4894-9f1f-30c37f9b0b19\") " Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.919640 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph" (OuterVolumeSpecName: "ceph") pod "f37d30c9-da2b-4894-9f1f-30c37f9b0b19" (UID: "f37d30c9-da2b-4894-9f1f-30c37f9b0b19"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.921342 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6" (OuterVolumeSpecName: "kube-api-access-pxbp6") pod "f37d30c9-da2b-4894-9f1f-30c37f9b0b19" (UID: "f37d30c9-da2b-4894-9f1f-30c37f9b0b19"). InnerVolumeSpecName "kube-api-access-pxbp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.949112 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f37d30c9-da2b-4894-9f1f-30c37f9b0b19" (UID: "f37d30c9-da2b-4894-9f1f-30c37f9b0b19"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:09 crc kubenswrapper[4959]: I1003 15:40:09.951157 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory" (OuterVolumeSpecName: "inventory") pod "f37d30c9-da2b-4894-9f1f-30c37f9b0b19" (UID: "f37d30c9-da2b-4894-9f1f-30c37f9b0b19"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.015405 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.015443 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxbp6\" (UniqueName: \"kubernetes.io/projected/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-kube-api-access-pxbp6\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.015460 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.015470 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f37d30c9-da2b-4894-9f1f-30c37f9b0b19-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.396489 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" event={"ID":"f37d30c9-da2b-4894-9f1f-30c37f9b0b19","Type":"ContainerDied","Data":"df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab"} Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.396774 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df9896380e6e419d88efaf6bbc5cafa2b8a85496ab5d0efd5b42e6a83c7964ab" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.396553 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-c9lbc" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.472739 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mtg8s"] Oct 03 15:40:10 crc kubenswrapper[4959]: E1003 15:40:10.473489 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f37d30c9-da2b-4894-9f1f-30c37f9b0b19" containerName="run-os-openstack-openstack-cell1" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.473599 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f37d30c9-da2b-4894-9f1f-30c37f9b0b19" containerName="run-os-openstack-openstack-cell1" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.473972 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f37d30c9-da2b-4894-9f1f-30c37f9b0b19" containerName="run-os-openstack-openstack-cell1" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.475927 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.487489 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.487502 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.487582 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.487750 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.511764 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mtg8s"] Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.627544 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.627599 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.627623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.627645 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59zjq\" (UniqueName: \"kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.729304 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.729359 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.729383 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59zjq\" (UniqueName: \"kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.729659 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.734997 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.735459 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.748071 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.757848 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59zjq\" (UniqueName: \"kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq\") pod \"reboot-os-openstack-openstack-cell1-mtg8s\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:10 crc kubenswrapper[4959]: I1003 15:40:10.800803 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:11 crc kubenswrapper[4959]: I1003 15:40:11.327924 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mtg8s"] Oct 03 15:40:11 crc kubenswrapper[4959]: I1003 15:40:11.407818 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" event={"ID":"4a88799d-3ffb-40b1-b444-0300adede7bc","Type":"ContainerStarted","Data":"2b360b0b6e41e3116e4545c2e851d5fee2035fac156eb6aeba3fb45b79e045a3"} Oct 03 15:40:12 crc kubenswrapper[4959]: I1003 15:40:12.424914 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" event={"ID":"4a88799d-3ffb-40b1-b444-0300adede7bc","Type":"ContainerStarted","Data":"cae48942ea4b2bafac563ab8ea842e55009f040e7bef04c9a0bdf01db2627cd0"} Oct 03 15:40:12 crc kubenswrapper[4959]: I1003 15:40:12.470248 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" podStartSLOduration=2.337831931 podStartE2EDuration="2.470225527s" podCreationTimestamp="2025-10-03 15:40:10 +0000 UTC" firstStartedPulling="2025-10-03 15:40:11.34616854 +0000 UTC m=+7780.549511977" lastFinishedPulling="2025-10-03 15:40:11.478562156 +0000 UTC m=+7780.681905573" observedRunningTime="2025-10-03 15:40:12.454585504 +0000 UTC m=+7781.657928961" watchObservedRunningTime="2025-10-03 15:40:12.470225527 +0000 UTC m=+7781.673568944" Oct 03 15:40:19 crc kubenswrapper[4959]: I1003 15:40:19.686078 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:40:19 crc kubenswrapper[4959]: E1003 15:40:19.687091 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:40:27 crc kubenswrapper[4959]: I1003 15:40:27.653426 4959 generic.go:334] "Generic (PLEG): container finished" podID="4a88799d-3ffb-40b1-b444-0300adede7bc" containerID="cae48942ea4b2bafac563ab8ea842e55009f040e7bef04c9a0bdf01db2627cd0" exitCode=0 Oct 03 15:40:27 crc kubenswrapper[4959]: I1003 15:40:27.653489 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" event={"ID":"4a88799d-3ffb-40b1-b444-0300adede7bc","Type":"ContainerDied","Data":"cae48942ea4b2bafac563ab8ea842e55009f040e7bef04c9a0bdf01db2627cd0"} Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.240705 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.305776 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory\") pod \"4a88799d-3ffb-40b1-b444-0300adede7bc\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.305863 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph\") pod \"4a88799d-3ffb-40b1-b444-0300adede7bc\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.305928 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59zjq\" (UniqueName: \"kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq\") pod \"4a88799d-3ffb-40b1-b444-0300adede7bc\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.305977 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key\") pod \"4a88799d-3ffb-40b1-b444-0300adede7bc\" (UID: \"4a88799d-3ffb-40b1-b444-0300adede7bc\") " Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.312312 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph" (OuterVolumeSpecName: "ceph") pod "4a88799d-3ffb-40b1-b444-0300adede7bc" (UID: "4a88799d-3ffb-40b1-b444-0300adede7bc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.314563 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq" (OuterVolumeSpecName: "kube-api-access-59zjq") pod "4a88799d-3ffb-40b1-b444-0300adede7bc" (UID: "4a88799d-3ffb-40b1-b444-0300adede7bc"). InnerVolumeSpecName "kube-api-access-59zjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.342956 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory" (OuterVolumeSpecName: "inventory") pod "4a88799d-3ffb-40b1-b444-0300adede7bc" (UID: "4a88799d-3ffb-40b1-b444-0300adede7bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.367453 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a88799d-3ffb-40b1-b444-0300adede7bc" (UID: "4a88799d-3ffb-40b1-b444-0300adede7bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.409528 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.409711 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.409825 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59zjq\" (UniqueName: \"kubernetes.io/projected/4a88799d-3ffb-40b1-b444-0300adede7bc-kube-api-access-59zjq\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.409933 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a88799d-3ffb-40b1-b444-0300adede7bc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.684161 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" event={"ID":"4a88799d-3ffb-40b1-b444-0300adede7bc","Type":"ContainerDied","Data":"2b360b0b6e41e3116e4545c2e851d5fee2035fac156eb6aeba3fb45b79e045a3"} Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.684264 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b360b0b6e41e3116e4545c2e851d5fee2035fac156eb6aeba3fb45b79e045a3" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.684265 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mtg8s" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.775520 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-w7ksg"] Oct 03 15:40:29 crc kubenswrapper[4959]: E1003 15:40:29.776022 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a88799d-3ffb-40b1-b444-0300adede7bc" containerName="reboot-os-openstack-openstack-cell1" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.776045 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a88799d-3ffb-40b1-b444-0300adede7bc" containerName="reboot-os-openstack-openstack-cell1" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.776330 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a88799d-3ffb-40b1-b444-0300adede7bc" containerName="reboot-os-openstack-openstack-cell1" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.777246 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.780665 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.780863 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.780931 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.781056 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.803525 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-w7ksg"] Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.842556 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.842651 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.842697 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.842826 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5g7p\" (UniqueName: \"kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.842888 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.843162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.843953 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.844001 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.844139 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.844251 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.844378 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.844515 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946275 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946331 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946372 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946416 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946470 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946541 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946577 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946597 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5g7p\" (UniqueName: \"kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946662 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946680 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.946711 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.953981 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.955016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.955156 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.955185 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.955469 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.955539 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.956223 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.956428 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.958532 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.959124 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.960634 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:29 crc kubenswrapper[4959]: I1003 15:40:29.966991 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5g7p\" (UniqueName: \"kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p\") pod \"install-certs-openstack-openstack-cell1-w7ksg\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:30 crc kubenswrapper[4959]: I1003 15:40:30.098384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:30 crc kubenswrapper[4959]: I1003 15:40:30.776215 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-w7ksg"] Oct 03 15:40:31 crc kubenswrapper[4959]: I1003 15:40:31.740289 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" event={"ID":"46ba8fc5-d9cf-4e16-aba1-b9145d389702","Type":"ContainerStarted","Data":"29b0235327dda3918a8c7e513de2dfb7aa008c05533a1a2e603f47ef185fa035"} Oct 03 15:40:31 crc kubenswrapper[4959]: I1003 15:40:31.740827 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" event={"ID":"46ba8fc5-d9cf-4e16-aba1-b9145d389702","Type":"ContainerStarted","Data":"4bd5e24bd1e41c5e9c3044d84b2ede4291afc5531a074d5664298a793d4e1638"} Oct 03 15:40:31 crc kubenswrapper[4959]: I1003 15:40:31.769619 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" podStartSLOduration=2.6197407459999997 podStartE2EDuration="2.769491057s" podCreationTimestamp="2025-10-03 15:40:29 +0000 UTC" firstStartedPulling="2025-10-03 15:40:30.767257648 +0000 UTC m=+7799.970601065" lastFinishedPulling="2025-10-03 15:40:30.917007959 +0000 UTC m=+7800.120351376" observedRunningTime="2025-10-03 15:40:31.766785351 +0000 UTC m=+7800.970128828" watchObservedRunningTime="2025-10-03 15:40:31.769491057 +0000 UTC m=+7800.972834474" Oct 03 15:40:32 crc kubenswrapper[4959]: I1003 15:40:32.686511 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:40:32 crc kubenswrapper[4959]: E1003 15:40:32.686906 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:40:44 crc kubenswrapper[4959]: I1003 15:40:44.686173 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:40:44 crc kubenswrapper[4959]: E1003 15:40:44.687287 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:40:50 crc kubenswrapper[4959]: I1003 15:40:50.964279 4959 generic.go:334] "Generic (PLEG): container finished" podID="46ba8fc5-d9cf-4e16-aba1-b9145d389702" containerID="29b0235327dda3918a8c7e513de2dfb7aa008c05533a1a2e603f47ef185fa035" exitCode=0 Oct 03 15:40:50 crc kubenswrapper[4959]: I1003 15:40:50.964350 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" event={"ID":"46ba8fc5-d9cf-4e16-aba1-b9145d389702","Type":"ContainerDied","Data":"29b0235327dda3918a8c7e513de2dfb7aa008c05533a1a2e603f47ef185fa035"} Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.467542 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.589641 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.589730 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.589793 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.589905 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.589946 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.590843 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.590887 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.590920 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.591010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.591091 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.591130 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5g7p\" (UniqueName: \"kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.591236 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph\") pod \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\" (UID: \"46ba8fc5-d9cf-4e16-aba1-b9145d389702\") " Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.597131 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.598147 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.598632 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.598861 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph" (OuterVolumeSpecName: "ceph") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.598942 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.599146 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.599241 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.599891 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.601234 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.602117 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p" (OuterVolumeSpecName: "kube-api-access-r5g7p") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "kube-api-access-r5g7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.626888 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory" (OuterVolumeSpecName: "inventory") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.640545 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46ba8fc5-d9cf-4e16-aba1-b9145d389702" (UID: "46ba8fc5-d9cf-4e16-aba1-b9145d389702"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696012 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696074 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696098 4959 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696116 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696136 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696154 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5g7p\" (UniqueName: \"kubernetes.io/projected/46ba8fc5-d9cf-4e16-aba1-b9145d389702-kube-api-access-r5g7p\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696169 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696183 4959 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696216 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696228 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696241 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.696254 4959 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46ba8fc5-d9cf-4e16-aba1-b9145d389702-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.985844 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" event={"ID":"46ba8fc5-d9cf-4e16-aba1-b9145d389702","Type":"ContainerDied","Data":"4bd5e24bd1e41c5e9c3044d84b2ede4291afc5531a074d5664298a793d4e1638"} Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.986252 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd5e24bd1e41c5e9c3044d84b2ede4291afc5531a074d5664298a793d4e1638" Oct 03 15:40:52 crc kubenswrapper[4959]: I1003 15:40:52.985947 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-w7ksg" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.104225 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-xp2jr"] Oct 03 15:40:53 crc kubenswrapper[4959]: E1003 15:40:53.104786 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ba8fc5-d9cf-4e16-aba1-b9145d389702" containerName="install-certs-openstack-openstack-cell1" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.104811 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ba8fc5-d9cf-4e16-aba1-b9145d389702" containerName="install-certs-openstack-openstack-cell1" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.105112 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ba8fc5-d9cf-4e16-aba1-b9145d389702" containerName="install-certs-openstack-openstack-cell1" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.106154 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.108913 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.108954 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.109100 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.111144 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.117450 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-xp2jr"] Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.209449 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.209650 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv5xv\" (UniqueName: \"kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.209753 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.210141 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.312400 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.312502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.312559 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv5xv\" (UniqueName: \"kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.312600 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.316409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.316783 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.323723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.330162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv5xv\" (UniqueName: \"kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv\") pod \"ceph-client-openstack-openstack-cell1-xp2jr\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:53 crc kubenswrapper[4959]: I1003 15:40:53.423928 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:40:54 crc kubenswrapper[4959]: I1003 15:40:54.002110 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-xp2jr"] Oct 03 15:40:55 crc kubenswrapper[4959]: I1003 15:40:55.022047 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" event={"ID":"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88","Type":"ContainerStarted","Data":"b39e5da4af43541dd0a7974a69d9e2a5feeaaea2dff28d0d97465da50524ce6f"} Oct 03 15:40:55 crc kubenswrapper[4959]: I1003 15:40:55.022654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" event={"ID":"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88","Type":"ContainerStarted","Data":"8522a1badc9dd833f030b6bc9fcb2870d05326fccc72fe5348e92a92b56816ef"} Oct 03 15:40:55 crc kubenswrapper[4959]: I1003 15:40:55.048548 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" podStartSLOduration=1.902472721 podStartE2EDuration="2.048526981s" podCreationTimestamp="2025-10-03 15:40:53 +0000 UTC" firstStartedPulling="2025-10-03 15:40:54.006827058 +0000 UTC m=+7823.210170515" lastFinishedPulling="2025-10-03 15:40:54.152881318 +0000 UTC m=+7823.356224775" observedRunningTime="2025-10-03 15:40:55.045476487 +0000 UTC m=+7824.248819974" watchObservedRunningTime="2025-10-03 15:40:55.048526981 +0000 UTC m=+7824.251870418" Oct 03 15:40:57 crc kubenswrapper[4959]: I1003 15:40:57.687288 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:40:57 crc kubenswrapper[4959]: E1003 15:40:57.688154 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:41:00 crc kubenswrapper[4959]: I1003 15:41:00.094014 4959 generic.go:334] "Generic (PLEG): container finished" podID="b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" containerID="b39e5da4af43541dd0a7974a69d9e2a5feeaaea2dff28d0d97465da50524ce6f" exitCode=0 Oct 03 15:41:00 crc kubenswrapper[4959]: I1003 15:41:00.094108 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" event={"ID":"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88","Type":"ContainerDied","Data":"b39e5da4af43541dd0a7974a69d9e2a5feeaaea2dff28d0d97465da50524ce6f"} Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.626033 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.821949 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory\") pod \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.822148 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv5xv\" (UniqueName: \"kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv\") pod \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.822331 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key\") pod \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.822413 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph\") pod \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\" (UID: \"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88\") " Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.830793 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph" (OuterVolumeSpecName: "ceph") pod "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" (UID: "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.831625 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv" (OuterVolumeSpecName: "kube-api-access-wv5xv") pod "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" (UID: "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88"). InnerVolumeSpecName "kube-api-access-wv5xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.857241 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" (UID: "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.860684 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory" (OuterVolumeSpecName: "inventory") pod "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" (UID: "b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.925929 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.925962 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv5xv\" (UniqueName: \"kubernetes.io/projected/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-kube-api-access-wv5xv\") on node \"crc\" DevicePath \"\"" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.925973 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:41:01 crc kubenswrapper[4959]: I1003 15:41:01.925981 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.126475 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" event={"ID":"b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88","Type":"ContainerDied","Data":"8522a1badc9dd833f030b6bc9fcb2870d05326fccc72fe5348e92a92b56816ef"} Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.127047 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8522a1badc9dd833f030b6bc9fcb2870d05326fccc72fe5348e92a92b56816ef" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.126615 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-xp2jr" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.235870 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-scpcf"] Oct 03 15:41:02 crc kubenswrapper[4959]: E1003 15:41:02.237613 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" containerName="ceph-client-openstack-openstack-cell1" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.237731 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" containerName="ceph-client-openstack-openstack-cell1" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.238073 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88" containerName="ceph-client-openstack-openstack-cell1" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.242217 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.245204 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.245455 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.245735 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.245829 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.248642 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.266591 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-scpcf"] Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.335830 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.336052 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.336203 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.336238 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czw9k\" (UniqueName: \"kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.336276 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.336322 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: E1003 15:41:02.393228 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a9268b_3fd9_4a2c_95b9_df1c8e67ea88.slice/crio-8522a1badc9dd833f030b6bc9fcb2870d05326fccc72fe5348e92a92b56816ef\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a9268b_3fd9_4a2c_95b9_df1c8e67ea88.slice\": RecentStats: unable to find data in memory cache]" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.438863 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.438914 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.439021 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.439044 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czw9k\" (UniqueName: \"kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.439075 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.439093 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.439890 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.445309 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.446896 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.451365 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.458362 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.459409 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czw9k\" (UniqueName: \"kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k\") pod \"ovn-openstack-openstack-cell1-scpcf\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:02 crc kubenswrapper[4959]: I1003 15:41:02.560377 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:41:03 crc kubenswrapper[4959]: I1003 15:41:03.100994 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-scpcf"] Oct 03 15:41:03 crc kubenswrapper[4959]: I1003 15:41:03.147001 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-scpcf" event={"ID":"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3","Type":"ContainerStarted","Data":"c8661d62289a62fc50fd41969baf1c62f3a25446ba3ab3b7d85c03ee8a28d3d2"} Oct 03 15:41:04 crc kubenswrapper[4959]: I1003 15:41:04.159794 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-scpcf" event={"ID":"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3","Type":"ContainerStarted","Data":"638053a77407ee658421633ef607e79d161cd1bb44cb24de84ba997dcb9b39b0"} Oct 03 15:41:04 crc kubenswrapper[4959]: I1003 15:41:04.191318 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-scpcf" podStartSLOduration=1.825353907 podStartE2EDuration="2.190877702s" podCreationTimestamp="2025-10-03 15:41:02 +0000 UTC" firstStartedPulling="2025-10-03 15:41:03.108861582 +0000 UTC m=+7832.312204989" lastFinishedPulling="2025-10-03 15:41:03.474385327 +0000 UTC m=+7832.677728784" observedRunningTime="2025-10-03 15:41:04.176159511 +0000 UTC m=+7833.379502928" watchObservedRunningTime="2025-10-03 15:41:04.190877702 +0000 UTC m=+7833.394221119" Oct 03 15:41:11 crc kubenswrapper[4959]: I1003 15:41:11.701417 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:41:11 crc kubenswrapper[4959]: E1003 15:41:11.702852 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:41:23 crc kubenswrapper[4959]: I1003 15:41:23.688164 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:41:23 crc kubenswrapper[4959]: E1003 15:41:23.689604 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:41:37 crc kubenswrapper[4959]: I1003 15:41:37.689240 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:41:37 crc kubenswrapper[4959]: E1003 15:41:37.690148 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:41:50 crc kubenswrapper[4959]: I1003 15:41:50.685630 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:41:50 crc kubenswrapper[4959]: E1003 15:41:50.686588 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:42:01 crc kubenswrapper[4959]: I1003 15:42:01.711370 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:42:01 crc kubenswrapper[4959]: E1003 15:42:01.712854 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:42:13 crc kubenswrapper[4959]: I1003 15:42:13.017354 4959 generic.go:334] "Generic (PLEG): container finished" podID="8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" containerID="638053a77407ee658421633ef607e79d161cd1bb44cb24de84ba997dcb9b39b0" exitCode=0 Oct 03 15:42:13 crc kubenswrapper[4959]: I1003 15:42:13.017438 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-scpcf" event={"ID":"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3","Type":"ContainerDied","Data":"638053a77407ee658421633ef607e79d161cd1bb44cb24de84ba997dcb9b39b0"} Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.632409 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779125 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779224 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czw9k\" (UniqueName: \"kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779251 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779283 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779360 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.779404 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle\") pod \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\" (UID: \"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3\") " Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.785244 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k" (OuterVolumeSpecName: "kube-api-access-czw9k") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "kube-api-access-czw9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.785906 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.800849 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph" (OuterVolumeSpecName: "ceph") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.807247 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.811999 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory" (OuterVolumeSpecName: "inventory") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.816135 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" (UID: "8df4e8ab-06f9-4a37-8e51-5954f63f2cf3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882229 4959 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882260 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czw9k\" (UniqueName: \"kubernetes.io/projected/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-kube-api-access-czw9k\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882269 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882777 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882787 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:14 crc kubenswrapper[4959]: I1003 15:42:14.882797 4959 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df4e8ab-06f9-4a37-8e51-5954f63f2cf3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.040656 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-scpcf" event={"ID":"8df4e8ab-06f9-4a37-8e51-5954f63f2cf3","Type":"ContainerDied","Data":"c8661d62289a62fc50fd41969baf1c62f3a25446ba3ab3b7d85c03ee8a28d3d2"} Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.041370 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8661d62289a62fc50fd41969baf1c62f3a25446ba3ab3b7d85c03ee8a28d3d2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.040974 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-scpcf" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.144991 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-877z2"] Oct 03 15:42:15 crc kubenswrapper[4959]: E1003 15:42:15.145766 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" containerName="ovn-openstack-openstack-cell1" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.145869 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" containerName="ovn-openstack-openstack-cell1" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.146233 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df4e8ab-06f9-4a37-8e51-5954f63f2cf3" containerName="ovn-openstack-openstack-cell1" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.147584 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.151663 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.151700 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.152022 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.152223 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.152291 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.152411 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.160791 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-877z2"] Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292066 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292119 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292506 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292661 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292741 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292779 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcxg2\" (UniqueName: \"kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.292976 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394380 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394453 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcxg2\" (UniqueName: \"kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394502 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394564 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394580 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.394638 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.398318 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.399451 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.401583 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.401663 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.402905 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.404430 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.422294 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcxg2\" (UniqueName: \"kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2\") pod \"neutron-metadata-openstack-openstack-cell1-877z2\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.476807 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:42:15 crc kubenswrapper[4959]: I1003 15:42:15.685572 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:42:15 crc kubenswrapper[4959]: E1003 15:42:15.686044 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:42:16 crc kubenswrapper[4959]: I1003 15:42:16.058949 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-877z2"] Oct 03 15:42:17 crc kubenswrapper[4959]: I1003 15:42:17.067682 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" event={"ID":"5f58accc-ca6d-4db4-b69c-200fb1c1e35e","Type":"ContainerStarted","Data":"a9d82d47d01a5b77059b2d164020634a0acc970d894a987af3238d3811a74d96"} Oct 03 15:42:17 crc kubenswrapper[4959]: I1003 15:42:17.068061 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" event={"ID":"5f58accc-ca6d-4db4-b69c-200fb1c1e35e","Type":"ContainerStarted","Data":"d7886766bfe02ea26f83ac7e1401a43edc7c368466deb9ad80081d803d45dd9f"} Oct 03 15:42:17 crc kubenswrapper[4959]: I1003 15:42:17.109291 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" podStartSLOduration=1.960604069 podStartE2EDuration="2.109271103s" podCreationTimestamp="2025-10-03 15:42:15 +0000 UTC" firstStartedPulling="2025-10-03 15:42:16.065999711 +0000 UTC m=+7905.269343148" lastFinishedPulling="2025-10-03 15:42:16.214666755 +0000 UTC m=+7905.418010182" observedRunningTime="2025-10-03 15:42:17.099916315 +0000 UTC m=+7906.303259742" watchObservedRunningTime="2025-10-03 15:42:17.109271103 +0000 UTC m=+7906.312614530" Oct 03 15:42:29 crc kubenswrapper[4959]: I1003 15:42:29.686496 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:42:29 crc kubenswrapper[4959]: E1003 15:42:29.687725 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:42:44 crc kubenswrapper[4959]: I1003 15:42:44.685434 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:42:44 crc kubenswrapper[4959]: E1003 15:42:44.686405 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:42:57 crc kubenswrapper[4959]: I1003 15:42:57.685531 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:42:57 crc kubenswrapper[4959]: E1003 15:42:57.686271 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:43:10 crc kubenswrapper[4959]: I1003 15:43:10.706573 4959 generic.go:334] "Generic (PLEG): container finished" podID="5f58accc-ca6d-4db4-b69c-200fb1c1e35e" containerID="a9d82d47d01a5b77059b2d164020634a0acc970d894a987af3238d3811a74d96" exitCode=0 Oct 03 15:43:10 crc kubenswrapper[4959]: I1003 15:43:10.707084 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" event={"ID":"5f58accc-ca6d-4db4-b69c-200fb1c1e35e","Type":"ContainerDied","Data":"a9d82d47d01a5b77059b2d164020634a0acc970d894a987af3238d3811a74d96"} Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.301528 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414317 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414455 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcxg2\" (UniqueName: \"kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414509 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414651 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414716 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414751 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.414862 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph\") pod \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\" (UID: \"5f58accc-ca6d-4db4-b69c-200fb1c1e35e\") " Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.420110 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2" (OuterVolumeSpecName: "kube-api-access-rcxg2") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "kube-api-access-rcxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.420348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.420474 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph" (OuterVolumeSpecName: "ceph") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.444267 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.447357 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.456253 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.459297 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory" (OuterVolumeSpecName: "inventory") pod "5f58accc-ca6d-4db4-b69c-200fb1c1e35e" (UID: "5f58accc-ca6d-4db4-b69c-200fb1c1e35e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.516984 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517014 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517026 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcxg2\" (UniqueName: \"kubernetes.io/projected/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-kube-api-access-rcxg2\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517036 4959 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517044 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517053 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.517063 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f58accc-ca6d-4db4-b69c-200fb1c1e35e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.686341 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:43:12 crc kubenswrapper[4959]: E1003 15:43:12.686618 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.729109 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" event={"ID":"5f58accc-ca6d-4db4-b69c-200fb1c1e35e","Type":"ContainerDied","Data":"d7886766bfe02ea26f83ac7e1401a43edc7c368466deb9ad80081d803d45dd9f"} Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.729147 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7886766bfe02ea26f83ac7e1401a43edc7c368466deb9ad80081d803d45dd9f" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.729172 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-877z2" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.818430 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nxgb8"] Oct 03 15:43:12 crc kubenswrapper[4959]: E1003 15:43:12.822299 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f58accc-ca6d-4db4-b69c-200fb1c1e35e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.822340 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f58accc-ca6d-4db4-b69c-200fb1c1e35e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.823253 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f58accc-ca6d-4db4-b69c-200fb1c1e35e" containerName="neutron-metadata-openstack-openstack-cell1" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.827272 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.831311 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.833149 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.833526 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.833941 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.835920 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.862838 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nxgb8"] Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933686 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933755 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933899 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933938 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:12 crc kubenswrapper[4959]: I1003 15:43:12.933971 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2vwt\" (UniqueName: \"kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035595 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035694 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035733 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035770 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2vwt\" (UniqueName: \"kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035868 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.035907 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.040398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.040913 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.041248 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.041718 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.042138 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.057313 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2vwt\" (UniqueName: \"kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt\") pod \"libvirt-openstack-openstack-cell1-nxgb8\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.159064 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.705596 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.707672 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-nxgb8"] Oct 03 15:43:13 crc kubenswrapper[4959]: I1003 15:43:13.740774 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" event={"ID":"86133afd-9abd-4209-bdc7-b729e26d8766","Type":"ContainerStarted","Data":"b2c53a603157eda43d4eb1d231e441e8f2d4426331e54ddb7ae9ca36a1a3b9ec"} Oct 03 15:43:14 crc kubenswrapper[4959]: I1003 15:43:14.749753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" event={"ID":"86133afd-9abd-4209-bdc7-b729e26d8766","Type":"ContainerStarted","Data":"eec9d75acb0a56c4e4ed3ed8e4ad4fb00490914470dc2cd7da952ff83c41a7b8"} Oct 03 15:43:23 crc kubenswrapper[4959]: I1003 15:43:23.686248 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:43:23 crc kubenswrapper[4959]: E1003 15:43:23.687219 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:43:37 crc kubenswrapper[4959]: I1003 15:43:37.686576 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:43:37 crc kubenswrapper[4959]: E1003 15:43:37.687736 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:43:51 crc kubenswrapper[4959]: I1003 15:43:51.705345 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:43:51 crc kubenswrapper[4959]: E1003 15:43:51.707138 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.084700 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" podStartSLOduration=41.872570096 podStartE2EDuration="42.084683568s" podCreationTimestamp="2025-10-03 15:43:12 +0000 UTC" firstStartedPulling="2025-10-03 15:43:13.705334069 +0000 UTC m=+7962.908677496" lastFinishedPulling="2025-10-03 15:43:13.917447541 +0000 UTC m=+7963.120790968" observedRunningTime="2025-10-03 15:43:14.777892482 +0000 UTC m=+7963.981235969" watchObservedRunningTime="2025-10-03 15:43:54.084683568 +0000 UTC m=+8003.288026985" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.092654 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.094975 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.109004 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.270217 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.271255 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97qj7\" (UniqueName: \"kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.271424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.373480 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97qj7\" (UniqueName: \"kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.373557 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.373630 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.374127 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.374238 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.395441 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97qj7\" (UniqueName: \"kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7\") pod \"community-operators-6nxf7\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.413745 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:43:54 crc kubenswrapper[4959]: I1003 15:43:54.949472 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:43:54 crc kubenswrapper[4959]: W1003 15:43:54.971244 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e61b080_ad94_4c0e_a957_6446027eb556.slice/crio-e5f12e7214e60007788c0767c8bd292061bc5861b759a3a05d1acf255dcd9e03 WatchSource:0}: Error finding container e5f12e7214e60007788c0767c8bd292061bc5861b759a3a05d1acf255dcd9e03: Status 404 returned error can't find the container with id e5f12e7214e60007788c0767c8bd292061bc5861b759a3a05d1acf255dcd9e03 Oct 03 15:43:55 crc kubenswrapper[4959]: I1003 15:43:55.274434 4959 generic.go:334] "Generic (PLEG): container finished" podID="5e61b080-ad94-4c0e-a957-6446027eb556" containerID="ee2e719250f20adfa0e0e2dd6bebfac6aaf4cf4d22f441c60dccbe33212ec8d3" exitCode=0 Oct 03 15:43:55 crc kubenswrapper[4959]: I1003 15:43:55.274508 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerDied","Data":"ee2e719250f20adfa0e0e2dd6bebfac6aaf4cf4d22f441c60dccbe33212ec8d3"} Oct 03 15:43:55 crc kubenswrapper[4959]: I1003 15:43:55.274540 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerStarted","Data":"e5f12e7214e60007788c0767c8bd292061bc5861b759a3a05d1acf255dcd9e03"} Oct 03 15:43:57 crc kubenswrapper[4959]: I1003 15:43:57.297058 4959 generic.go:334] "Generic (PLEG): container finished" podID="5e61b080-ad94-4c0e-a957-6446027eb556" containerID="d987d3664930243670cc2ef1976c073ed4423e150d076440f42f348c9ee76b82" exitCode=0 Oct 03 15:43:57 crc kubenswrapper[4959]: I1003 15:43:57.297117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerDied","Data":"d987d3664930243670cc2ef1976c073ed4423e150d076440f42f348c9ee76b82"} Oct 03 15:43:58 crc kubenswrapper[4959]: I1003 15:43:58.314775 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerStarted","Data":"5cbf35973c91627d0a1d186470d3d7e0ec71557526d9e936844ac6a548c0505b"} Oct 03 15:43:58 crc kubenswrapper[4959]: I1003 15:43:58.349801 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nxf7" podStartSLOduration=1.922565509 podStartE2EDuration="4.349775669s" podCreationTimestamp="2025-10-03 15:43:54 +0000 UTC" firstStartedPulling="2025-10-03 15:43:55.279681582 +0000 UTC m=+8004.483024999" lastFinishedPulling="2025-10-03 15:43:57.706891742 +0000 UTC m=+8006.910235159" observedRunningTime="2025-10-03 15:43:58.338487013 +0000 UTC m=+8007.541830470" watchObservedRunningTime="2025-10-03 15:43:58.349775669 +0000 UTC m=+8007.553119096" Oct 03 15:44:04 crc kubenswrapper[4959]: I1003 15:44:04.414112 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:04 crc kubenswrapper[4959]: I1003 15:44:04.414826 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:04 crc kubenswrapper[4959]: I1003 15:44:04.479429 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:05 crc kubenswrapper[4959]: I1003 15:44:05.445925 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:05 crc kubenswrapper[4959]: I1003 15:44:05.503402 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:44:05 crc kubenswrapper[4959]: I1003 15:44:05.685380 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:44:05 crc kubenswrapper[4959]: E1003 15:44:05.685687 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:44:07 crc kubenswrapper[4959]: I1003 15:44:07.416853 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nxf7" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="registry-server" containerID="cri-o://5cbf35973c91627d0a1d186470d3d7e0ec71557526d9e936844ac6a548c0505b" gracePeriod=2 Oct 03 15:44:08 crc kubenswrapper[4959]: I1003 15:44:08.432793 4959 generic.go:334] "Generic (PLEG): container finished" podID="5e61b080-ad94-4c0e-a957-6446027eb556" containerID="5cbf35973c91627d0a1d186470d3d7e0ec71557526d9e936844ac6a548c0505b" exitCode=0 Oct 03 15:44:08 crc kubenswrapper[4959]: I1003 15:44:08.433172 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerDied","Data":"5cbf35973c91627d0a1d186470d3d7e0ec71557526d9e936844ac6a548c0505b"} Oct 03 15:44:08 crc kubenswrapper[4959]: I1003 15:44:08.997659 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.116029 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content\") pod \"5e61b080-ad94-4c0e-a957-6446027eb556\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.116260 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities\") pod \"5e61b080-ad94-4c0e-a957-6446027eb556\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.116309 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97qj7\" (UniqueName: \"kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7\") pod \"5e61b080-ad94-4c0e-a957-6446027eb556\" (UID: \"5e61b080-ad94-4c0e-a957-6446027eb556\") " Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.117933 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities" (OuterVolumeSpecName: "utilities") pod "5e61b080-ad94-4c0e-a957-6446027eb556" (UID: "5e61b080-ad94-4c0e-a957-6446027eb556"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.135923 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7" (OuterVolumeSpecName: "kube-api-access-97qj7") pod "5e61b080-ad94-4c0e-a957-6446027eb556" (UID: "5e61b080-ad94-4c0e-a957-6446027eb556"). InnerVolumeSpecName "kube-api-access-97qj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.172243 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e61b080-ad94-4c0e-a957-6446027eb556" (UID: "5e61b080-ad94-4c0e-a957-6446027eb556"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.219735 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.219863 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e61b080-ad94-4c0e-a957-6446027eb556-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.219895 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97qj7\" (UniqueName: \"kubernetes.io/projected/5e61b080-ad94-4c0e-a957-6446027eb556-kube-api-access-97qj7\") on node \"crc\" DevicePath \"\"" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.452027 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nxf7" event={"ID":"5e61b080-ad94-4c0e-a957-6446027eb556","Type":"ContainerDied","Data":"e5f12e7214e60007788c0767c8bd292061bc5861b759a3a05d1acf255dcd9e03"} Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.452141 4959 scope.go:117] "RemoveContainer" containerID="5cbf35973c91627d0a1d186470d3d7e0ec71557526d9e936844ac6a548c0505b" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.453426 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nxf7" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.491400 4959 scope.go:117] "RemoveContainer" containerID="d987d3664930243670cc2ef1976c073ed4423e150d076440f42f348c9ee76b82" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.506850 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.525253 4959 scope.go:117] "RemoveContainer" containerID="ee2e719250f20adfa0e0e2dd6bebfac6aaf4cf4d22f441c60dccbe33212ec8d3" Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.528825 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nxf7"] Oct 03 15:44:09 crc kubenswrapper[4959]: I1003 15:44:09.700789 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" path="/var/lib/kubelet/pods/5e61b080-ad94-4c0e-a957-6446027eb556/volumes" Oct 03 15:44:16 crc kubenswrapper[4959]: I1003 15:44:16.686185 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:44:16 crc kubenswrapper[4959]: E1003 15:44:16.687720 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:44:30 crc kubenswrapper[4959]: I1003 15:44:30.685709 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:44:30 crc kubenswrapper[4959]: E1003 15:44:30.686854 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.090579 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:44:42 crc kubenswrapper[4959]: E1003 15:44:42.091895 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="extract-content" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.091915 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="extract-content" Oct 03 15:44:42 crc kubenswrapper[4959]: E1003 15:44:42.091934 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="registry-server" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.091943 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="registry-server" Oct 03 15:44:42 crc kubenswrapper[4959]: E1003 15:44:42.091974 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="extract-utilities" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.091983 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="extract-utilities" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.092759 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e61b080-ad94-4c0e-a957-6446027eb556" containerName="registry-server" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.099541 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.103347 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.239880 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.240214 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.240358 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrwq\" (UniqueName: \"kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.342482 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.342565 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.342711 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrwq\" (UniqueName: \"kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.343358 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.343371 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.366372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrwq\" (UniqueName: \"kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq\") pod \"redhat-operators-gtfsj\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.430390 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.687058 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:44:42 crc kubenswrapper[4959]: I1003 15:44:42.947867 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:44:43 crc kubenswrapper[4959]: I1003 15:44:43.870334 4959 generic.go:334] "Generic (PLEG): container finished" podID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerID="4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6" exitCode=0 Oct 03 15:44:43 crc kubenswrapper[4959]: I1003 15:44:43.870486 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerDied","Data":"4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6"} Oct 03 15:44:43 crc kubenswrapper[4959]: I1003 15:44:43.870958 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerStarted","Data":"b21eaf5b1cb7f26deac380b3210e28e41b3f22862a6ded6599326554c75bf76c"} Oct 03 15:44:43 crc kubenswrapper[4959]: I1003 15:44:43.874543 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d"} Oct 03 15:44:44 crc kubenswrapper[4959]: I1003 15:44:44.887349 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerStarted","Data":"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e"} Oct 03 15:44:45 crc kubenswrapper[4959]: I1003 15:44:45.900962 4959 generic.go:334] "Generic (PLEG): container finished" podID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerID="dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e" exitCode=0 Oct 03 15:44:45 crc kubenswrapper[4959]: I1003 15:44:45.901132 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerDied","Data":"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e"} Oct 03 15:44:47 crc kubenswrapper[4959]: I1003 15:44:47.920100 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerStarted","Data":"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb"} Oct 03 15:44:47 crc kubenswrapper[4959]: I1003 15:44:47.937153 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gtfsj" podStartSLOduration=2.3468536540000002 podStartE2EDuration="5.937135648s" podCreationTimestamp="2025-10-03 15:44:42 +0000 UTC" firstStartedPulling="2025-10-03 15:44:43.872334611 +0000 UTC m=+8053.075678028" lastFinishedPulling="2025-10-03 15:44:47.462616595 +0000 UTC m=+8056.665960022" observedRunningTime="2025-10-03 15:44:47.934185726 +0000 UTC m=+8057.137529163" watchObservedRunningTime="2025-10-03 15:44:47.937135648 +0000 UTC m=+8057.140479065" Oct 03 15:44:52 crc kubenswrapper[4959]: I1003 15:44:52.430509 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:52 crc kubenswrapper[4959]: I1003 15:44:52.431061 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:44:53 crc kubenswrapper[4959]: I1003 15:44:53.479686 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gtfsj" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="registry-server" probeResult="failure" output=< Oct 03 15:44:53 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:44:53 crc kubenswrapper[4959]: > Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.152759 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8"] Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.155216 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.157309 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.157756 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.168093 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8"] Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.264487 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rczt\" (UniqueName: \"kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.264927 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.265091 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.368469 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rczt\" (UniqueName: \"kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.368586 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.368617 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.369528 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.382554 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.384795 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rczt\" (UniqueName: \"kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt\") pod \"collect-profiles-29325105-vsth8\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:00 crc kubenswrapper[4959]: I1003 15:45:00.487518 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:01 crc kubenswrapper[4959]: I1003 15:45:01.000746 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8"] Oct 03 15:45:01 crc kubenswrapper[4959]: W1003 15:45:01.005529 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde58d4f6_9bb0_46bc_aa24_5ff7cddb0b9f.slice/crio-8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db WatchSource:0}: Error finding container 8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db: Status 404 returned error can't find the container with id 8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db Oct 03 15:45:01 crc kubenswrapper[4959]: I1003 15:45:01.093638 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" event={"ID":"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f","Type":"ContainerStarted","Data":"8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db"} Oct 03 15:45:02 crc kubenswrapper[4959]: I1003 15:45:02.112928 4959 generic.go:334] "Generic (PLEG): container finished" podID="de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" containerID="0bbe108b71cf548f0828ea5f7977203d02b036d8b01408dbe7c0f80f484df00c" exitCode=0 Oct 03 15:45:02 crc kubenswrapper[4959]: I1003 15:45:02.114390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" event={"ID":"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f","Type":"ContainerDied","Data":"0bbe108b71cf548f0828ea5f7977203d02b036d8b01408dbe7c0f80f484df00c"} Oct 03 15:45:02 crc kubenswrapper[4959]: I1003 15:45:02.499978 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:45:02 crc kubenswrapper[4959]: I1003 15:45:02.565991 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.593779 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.642067 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rczt\" (UniqueName: \"kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt\") pod \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.642143 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume\") pod \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.642412 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume\") pod \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\" (UID: \"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f\") " Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.642961 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" (UID: "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.647819 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" (UID: "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.650486 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt" (OuterVolumeSpecName: "kube-api-access-7rczt") pod "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" (UID: "de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f"). InnerVolumeSpecName "kube-api-access-7rczt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.745826 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.746134 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rczt\" (UniqueName: \"kubernetes.io/projected/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-kube-api-access-7rczt\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:03 crc kubenswrapper[4959]: I1003 15:45:03.746150 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:04 crc kubenswrapper[4959]: I1003 15:45:04.139971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" event={"ID":"de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f","Type":"ContainerDied","Data":"8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db"} Oct 03 15:45:04 crc kubenswrapper[4959]: I1003 15:45:04.140019 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8240030e4748aae033d78d32f4a09b987608f015c6dd07803555a05c562aa7db" Oct 03 15:45:04 crc kubenswrapper[4959]: I1003 15:45:04.140052 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325105-vsth8" Oct 03 15:45:04 crc kubenswrapper[4959]: I1003 15:45:04.679552 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x"] Oct 03 15:45:04 crc kubenswrapper[4959]: I1003 15:45:04.701397 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325060-crq9x"] Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.278412 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.278687 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gtfsj" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="registry-server" containerID="cri-o://579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb" gracePeriod=2 Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.707975 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f4f4059-389d-4e04-90b3-34ec7a502754" path="/var/lib/kubelet/pods/6f4f4059-389d-4e04-90b3-34ec7a502754/volumes" Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.759093 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.899839 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content\") pod \"f055258c-ea6c-4983-b2f0-7b18667b934e\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.900333 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities\") pod \"f055258c-ea6c-4983-b2f0-7b18667b934e\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.900595 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrwq\" (UniqueName: \"kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq\") pod \"f055258c-ea6c-4983-b2f0-7b18667b934e\" (UID: \"f055258c-ea6c-4983-b2f0-7b18667b934e\") " Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.901313 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities" (OuterVolumeSpecName: "utilities") pod "f055258c-ea6c-4983-b2f0-7b18667b934e" (UID: "f055258c-ea6c-4983-b2f0-7b18667b934e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.901534 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.909401 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq" (OuterVolumeSpecName: "kube-api-access-fzrwq") pod "f055258c-ea6c-4983-b2f0-7b18667b934e" (UID: "f055258c-ea6c-4983-b2f0-7b18667b934e"). InnerVolumeSpecName "kube-api-access-fzrwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:45:05 crc kubenswrapper[4959]: I1003 15:45:05.981872 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f055258c-ea6c-4983-b2f0-7b18667b934e" (UID: "f055258c-ea6c-4983-b2f0-7b18667b934e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.002864 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrwq\" (UniqueName: \"kubernetes.io/projected/f055258c-ea6c-4983-b2f0-7b18667b934e-kube-api-access-fzrwq\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.002931 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f055258c-ea6c-4983-b2f0-7b18667b934e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.167467 4959 generic.go:334] "Generic (PLEG): container finished" podID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerID="579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb" exitCode=0 Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.167509 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerDied","Data":"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb"} Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.167535 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtfsj" event={"ID":"f055258c-ea6c-4983-b2f0-7b18667b934e","Type":"ContainerDied","Data":"b21eaf5b1cb7f26deac380b3210e28e41b3f22862a6ded6599326554c75bf76c"} Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.167551 4959 scope.go:117] "RemoveContainer" containerID="579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.167697 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtfsj" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.192332 4959 scope.go:117] "RemoveContainer" containerID="dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.210341 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.219840 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gtfsj"] Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.238958 4959 scope.go:117] "RemoveContainer" containerID="4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.270460 4959 scope.go:117] "RemoveContainer" containerID="579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb" Oct 03 15:45:06 crc kubenswrapper[4959]: E1003 15:45:06.270932 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb\": container with ID starting with 579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb not found: ID does not exist" containerID="579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.270961 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb"} err="failed to get container status \"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb\": rpc error: code = NotFound desc = could not find container \"579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb\": container with ID starting with 579f81d38fb5f3c26a7f8ed4ff7a9c2ef3e830235b274f196a49ad33ed24eccb not found: ID does not exist" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.270987 4959 scope.go:117] "RemoveContainer" containerID="dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e" Oct 03 15:45:06 crc kubenswrapper[4959]: E1003 15:45:06.271287 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e\": container with ID starting with dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e not found: ID does not exist" containerID="dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.271311 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e"} err="failed to get container status \"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e\": rpc error: code = NotFound desc = could not find container \"dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e\": container with ID starting with dc4d4ae7a65d26383ae8479ef9158971f07e3d0b51207fa1a9156ddb16040a7e not found: ID does not exist" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.271327 4959 scope.go:117] "RemoveContainer" containerID="4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6" Oct 03 15:45:06 crc kubenswrapper[4959]: E1003 15:45:06.271592 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6\": container with ID starting with 4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6 not found: ID does not exist" containerID="4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6" Oct 03 15:45:06 crc kubenswrapper[4959]: I1003 15:45:06.271613 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6"} err="failed to get container status \"4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6\": rpc error: code = NotFound desc = could not find container \"4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6\": container with ID starting with 4616e9a49b0eb2e8362d7ee978cfe0b87feda23e1753a1e65139a16b34a154a6 not found: ID does not exist" Oct 03 15:45:07 crc kubenswrapper[4959]: I1003 15:45:07.701967 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" path="/var/lib/kubelet/pods/f055258c-ea6c-4983-b2f0-7b18667b934e/volumes" Oct 03 15:45:12 crc kubenswrapper[4959]: I1003 15:45:12.227851 4959 scope.go:117] "RemoveContainer" containerID="ccffb66f52a9bcc89ceb12109241ee53fea8bce5817981398ea259b975a5709e" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.470289 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:43 crc kubenswrapper[4959]: E1003 15:46:43.472230 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" containerName="collect-profiles" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472252 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" containerName="collect-profiles" Oct 03 15:46:43 crc kubenswrapper[4959]: E1003 15:46:43.472276 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="extract-utilities" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472283 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="extract-utilities" Oct 03 15:46:43 crc kubenswrapper[4959]: E1003 15:46:43.472295 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="registry-server" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472301 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="registry-server" Oct 03 15:46:43 crc kubenswrapper[4959]: E1003 15:46:43.472316 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="extract-content" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472322 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="extract-content" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472523 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f055258c-ea6c-4983-b2f0-7b18667b934e" containerName="registry-server" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.472535 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="de58d4f6-9bb0-46bc-aa24-5ff7cddb0b9f" containerName="collect-profiles" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.477249 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.517255 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.619353 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.619424 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsnq8\" (UniqueName: \"kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.619954 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.721703 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsnq8\" (UniqueName: \"kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.721839 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.721933 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.722372 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.722406 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.745907 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsnq8\" (UniqueName: \"kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8\") pod \"certified-operators-6s68c\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:43 crc kubenswrapper[4959]: I1003 15:46:43.822901 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:44 crc kubenswrapper[4959]: I1003 15:46:44.362148 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:45 crc kubenswrapper[4959]: I1003 15:46:45.372971 4959 generic.go:334] "Generic (PLEG): container finished" podID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerID="38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a" exitCode=0 Oct 03 15:46:45 crc kubenswrapper[4959]: I1003 15:46:45.373056 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerDied","Data":"38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a"} Oct 03 15:46:45 crc kubenswrapper[4959]: I1003 15:46:45.373327 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerStarted","Data":"66d2fc5874ccf947fcf2d6129b17b23a85c987951d36e86a5d343447b5b8f4e8"} Oct 03 15:46:47 crc kubenswrapper[4959]: I1003 15:46:47.396578 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerStarted","Data":"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1"} Oct 03 15:46:48 crc kubenswrapper[4959]: I1003 15:46:48.411822 4959 generic.go:334] "Generic (PLEG): container finished" podID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerID="0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1" exitCode=0 Oct 03 15:46:48 crc kubenswrapper[4959]: I1003 15:46:48.411926 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerDied","Data":"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1"} Oct 03 15:46:50 crc kubenswrapper[4959]: I1003 15:46:50.437144 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerStarted","Data":"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd"} Oct 03 15:46:50 crc kubenswrapper[4959]: I1003 15:46:50.461865 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6s68c" podStartSLOduration=3.652454035 podStartE2EDuration="7.461843792s" podCreationTimestamp="2025-10-03 15:46:43 +0000 UTC" firstStartedPulling="2025-10-03 15:46:45.375726773 +0000 UTC m=+8174.579070200" lastFinishedPulling="2025-10-03 15:46:49.18511652 +0000 UTC m=+8178.388459957" observedRunningTime="2025-10-03 15:46:50.458649763 +0000 UTC m=+8179.661993190" watchObservedRunningTime="2025-10-03 15:46:50.461843792 +0000 UTC m=+8179.665187229" Oct 03 15:46:53 crc kubenswrapper[4959]: I1003 15:46:53.824167 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:53 crc kubenswrapper[4959]: I1003 15:46:53.824855 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:53 crc kubenswrapper[4959]: I1003 15:46:53.880476 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:54 crc kubenswrapper[4959]: I1003 15:46:54.578493 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:54 crc kubenswrapper[4959]: I1003 15:46:54.649628 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:56 crc kubenswrapper[4959]: I1003 15:46:56.512887 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6s68c" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="registry-server" containerID="cri-o://1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd" gracePeriod=2 Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.049506 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.140103 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities\") pod \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.140170 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content\") pod \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.140204 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsnq8\" (UniqueName: \"kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8\") pod \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\" (UID: \"80d5528f-b454-4d9a-a83f-1fc5abbce8e1\") " Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.157786 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities" (OuterVolumeSpecName: "utilities") pod "80d5528f-b454-4d9a-a83f-1fc5abbce8e1" (UID: "80d5528f-b454-4d9a-a83f-1fc5abbce8e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.162479 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8" (OuterVolumeSpecName: "kube-api-access-nsnq8") pod "80d5528f-b454-4d9a-a83f-1fc5abbce8e1" (UID: "80d5528f-b454-4d9a-a83f-1fc5abbce8e1"). InnerVolumeSpecName "kube-api-access-nsnq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.221232 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80d5528f-b454-4d9a-a83f-1fc5abbce8e1" (UID: "80d5528f-b454-4d9a-a83f-1fc5abbce8e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.242395 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsnq8\" (UniqueName: \"kubernetes.io/projected/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-kube-api-access-nsnq8\") on node \"crc\" DevicePath \"\"" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.242425 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.242436 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80d5528f-b454-4d9a-a83f-1fc5abbce8e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.524141 4959 generic.go:334] "Generic (PLEG): container finished" podID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerID="1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd" exitCode=0 Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.524182 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerDied","Data":"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd"} Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.524231 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s68c" event={"ID":"80d5528f-b454-4d9a-a83f-1fc5abbce8e1","Type":"ContainerDied","Data":"66d2fc5874ccf947fcf2d6129b17b23a85c987951d36e86a5d343447b5b8f4e8"} Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.524203 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s68c" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.524254 4959 scope.go:117] "RemoveContainer" containerID="1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.557942 4959 scope.go:117] "RemoveContainer" containerID="0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.561287 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.575306 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6s68c"] Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.591213 4959 scope.go:117] "RemoveContainer" containerID="38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.652613 4959 scope.go:117] "RemoveContainer" containerID="1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd" Oct 03 15:46:57 crc kubenswrapper[4959]: E1003 15:46:57.653269 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd\": container with ID starting with 1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd not found: ID does not exist" containerID="1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.653307 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd"} err="failed to get container status \"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd\": rpc error: code = NotFound desc = could not find container \"1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd\": container with ID starting with 1778a53c1438c13b4677e5980887c824e28a930be34c92b4858adb47349131dd not found: ID does not exist" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.653335 4959 scope.go:117] "RemoveContainer" containerID="0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1" Oct 03 15:46:57 crc kubenswrapper[4959]: E1003 15:46:57.653681 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1\": container with ID starting with 0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1 not found: ID does not exist" containerID="0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.653706 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1"} err="failed to get container status \"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1\": rpc error: code = NotFound desc = could not find container \"0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1\": container with ID starting with 0333088ece0218aad317c0310d110e08d491c1866c0eb44655ec19c84ffc7ef1 not found: ID does not exist" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.653721 4959 scope.go:117] "RemoveContainer" containerID="38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a" Oct 03 15:46:57 crc kubenswrapper[4959]: E1003 15:46:57.654071 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a\": container with ID starting with 38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a not found: ID does not exist" containerID="38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.654106 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a"} err="failed to get container status \"38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a\": rpc error: code = NotFound desc = could not find container \"38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a\": container with ID starting with 38f1b679ad27d37e3fa5467bfc3e89447374e64417d174eed53c3100ec24ee9a not found: ID does not exist" Oct 03 15:46:57 crc kubenswrapper[4959]: I1003 15:46:57.698418 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" path="/var/lib/kubelet/pods/80d5528f-b454-4d9a-a83f-1fc5abbce8e1/volumes" Oct 03 15:47:06 crc kubenswrapper[4959]: I1003 15:47:06.044234 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:47:06 crc kubenswrapper[4959]: I1003 15:47:06.045073 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:47:36 crc kubenswrapper[4959]: I1003 15:47:36.047247 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:47:36 crc kubenswrapper[4959]: I1003 15:47:36.047985 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:47:54 crc kubenswrapper[4959]: I1003 15:47:54.175395 4959 generic.go:334] "Generic (PLEG): container finished" podID="86133afd-9abd-4209-bdc7-b729e26d8766" containerID="eec9d75acb0a56c4e4ed3ed8e4ad4fb00490914470dc2cd7da952ff83c41a7b8" exitCode=0 Oct 03 15:47:54 crc kubenswrapper[4959]: I1003 15:47:54.175584 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" event={"ID":"86133afd-9abd-4209-bdc7-b729e26d8766","Type":"ContainerDied","Data":"eec9d75acb0a56c4e4ed3ed8e4ad4fb00490914470dc2cd7da952ff83c41a7b8"} Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.691290 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.747919 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.747976 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.748020 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.748062 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2vwt\" (UniqueName: \"kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.748147 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.748328 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key\") pod \"86133afd-9abd-4209-bdc7-b729e26d8766\" (UID: \"86133afd-9abd-4209-bdc7-b729e26d8766\") " Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.753593 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph" (OuterVolumeSpecName: "ceph") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.766288 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.768025 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt" (OuterVolumeSpecName: "kube-api-access-k2vwt") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "kube-api-access-k2vwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.784574 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.786042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.787029 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory" (OuterVolumeSpecName: "inventory") pod "86133afd-9abd-4209-bdc7-b729e26d8766" (UID: "86133afd-9abd-4209-bdc7-b729e26d8766"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851852 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851891 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851905 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851919 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851931 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2vwt\" (UniqueName: \"kubernetes.io/projected/86133afd-9abd-4209-bdc7-b729e26d8766-kube-api-access-k2vwt\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:55 crc kubenswrapper[4959]: I1003 15:47:55.851942 4959 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86133afd-9abd-4209-bdc7-b729e26d8766-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.204358 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" event={"ID":"86133afd-9abd-4209-bdc7-b729e26d8766","Type":"ContainerDied","Data":"b2c53a603157eda43d4eb1d231e441e8f2d4426331e54ddb7ae9ca36a1a3b9ec"} Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.204651 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2c53a603157eda43d4eb1d231e441e8f2d4426331e54ddb7ae9ca36a1a3b9ec" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.204443 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-nxgb8" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.299591 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8w98r"] Oct 03 15:47:56 crc kubenswrapper[4959]: E1003 15:47:56.300300 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="extract-content" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.300375 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="extract-content" Oct 03 15:47:56 crc kubenswrapper[4959]: E1003 15:47:56.300435 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="registry-server" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.300487 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="registry-server" Oct 03 15:47:56 crc kubenswrapper[4959]: E1003 15:47:56.300582 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86133afd-9abd-4209-bdc7-b729e26d8766" containerName="libvirt-openstack-openstack-cell1" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.300634 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="86133afd-9abd-4209-bdc7-b729e26d8766" containerName="libvirt-openstack-openstack-cell1" Oct 03 15:47:56 crc kubenswrapper[4959]: E1003 15:47:56.300687 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="extract-utilities" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.300735 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="extract-utilities" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.300970 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="80d5528f-b454-4d9a-a83f-1fc5abbce8e1" containerName="registry-server" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.301041 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="86133afd-9abd-4209-bdc7-b729e26d8766" containerName="libvirt-openstack-openstack-cell1" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.301884 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.305031 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.306118 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.306163 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.307975 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.308016 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.308183 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.308324 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.328908 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8w98r"] Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362506 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362556 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362578 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362605 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362627 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362794 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362948 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.362975 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.363141 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.363165 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbf6m\" (UniqueName: \"kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.363183 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465469 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465533 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465624 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465650 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbf6m\" (UniqueName: \"kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465679 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465754 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465776 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465805 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465836 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465866 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.465920 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.467010 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.467791 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.472447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.472524 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.473000 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.473433 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.474692 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.475131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.476679 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.485790 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.489052 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbf6m\" (UniqueName: \"kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m\") pod \"nova-cell1-openstack-openstack-cell1-8w98r\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:56 crc kubenswrapper[4959]: I1003 15:47:56.621715 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.231376 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8w98r"] Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.675784 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.679765 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.715981 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.793993 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.794063 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.794151 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl6dc\" (UniqueName: \"kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.907068 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.907706 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.907729 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.908158 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl6dc\" (UniqueName: \"kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.908286 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:57 crc kubenswrapper[4959]: I1003 15:47:57.928085 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl6dc\" (UniqueName: \"kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc\") pod \"redhat-marketplace-7lxxq\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:58 crc kubenswrapper[4959]: I1003 15:47:58.008604 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:47:58 crc kubenswrapper[4959]: I1003 15:47:58.227117 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" event={"ID":"2a09699c-f954-45ac-b96b-accef4753ab3","Type":"ContainerStarted","Data":"1425b6f659d747820a2da6575cd116fdcfbcc7774839c0db2fafcc1875abb4a1"} Oct 03 15:47:58 crc kubenswrapper[4959]: I1003 15:47:58.227414 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" event={"ID":"2a09699c-f954-45ac-b96b-accef4753ab3","Type":"ContainerStarted","Data":"14acbd158741e1a5138328bff85654d54628c0b857b697177460d0c0511b8506"} Oct 03 15:47:58 crc kubenswrapper[4959]: I1003 15:47:58.251658 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" podStartSLOduration=2.038325835 podStartE2EDuration="2.251639916s" podCreationTimestamp="2025-10-03 15:47:56 +0000 UTC" firstStartedPulling="2025-10-03 15:47:57.237889119 +0000 UTC m=+8246.441232526" lastFinishedPulling="2025-10-03 15:47:57.45120319 +0000 UTC m=+8246.654546607" observedRunningTime="2025-10-03 15:47:58.244909801 +0000 UTC m=+8247.448253218" watchObservedRunningTime="2025-10-03 15:47:58.251639916 +0000 UTC m=+8247.454983323" Oct 03 15:47:58 crc kubenswrapper[4959]: I1003 15:47:58.505825 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:47:58 crc kubenswrapper[4959]: W1003 15:47:58.510539 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda348127e_b073_4ce9_b496_b36c0d704c6f.slice/crio-15fb442af3f8625769582423811958997d7a01679019f78fbadb43b302755264 WatchSource:0}: Error finding container 15fb442af3f8625769582423811958997d7a01679019f78fbadb43b302755264: Status 404 returned error can't find the container with id 15fb442af3f8625769582423811958997d7a01679019f78fbadb43b302755264 Oct 03 15:47:59 crc kubenswrapper[4959]: I1003 15:47:59.241874 4959 generic.go:334] "Generic (PLEG): container finished" podID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerID="50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30" exitCode=0 Oct 03 15:47:59 crc kubenswrapper[4959]: I1003 15:47:59.241971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerDied","Data":"50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30"} Oct 03 15:47:59 crc kubenswrapper[4959]: I1003 15:47:59.242302 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerStarted","Data":"15fb442af3f8625769582423811958997d7a01679019f78fbadb43b302755264"} Oct 03 15:48:00 crc kubenswrapper[4959]: I1003 15:48:00.252797 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerStarted","Data":"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427"} Oct 03 15:48:01 crc kubenswrapper[4959]: I1003 15:48:01.268752 4959 generic.go:334] "Generic (PLEG): container finished" podID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerID="473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427" exitCode=0 Oct 03 15:48:01 crc kubenswrapper[4959]: I1003 15:48:01.268868 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerDied","Data":"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427"} Oct 03 15:48:02 crc kubenswrapper[4959]: I1003 15:48:02.281317 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerStarted","Data":"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c"} Oct 03 15:48:02 crc kubenswrapper[4959]: I1003 15:48:02.309928 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7lxxq" podStartSLOduration=2.8099115550000002 podStartE2EDuration="5.309905803s" podCreationTimestamp="2025-10-03 15:47:57 +0000 UTC" firstStartedPulling="2025-10-03 15:47:59.245076926 +0000 UTC m=+8248.448420343" lastFinishedPulling="2025-10-03 15:48:01.745071154 +0000 UTC m=+8250.948414591" observedRunningTime="2025-10-03 15:48:02.302052432 +0000 UTC m=+8251.505395849" watchObservedRunningTime="2025-10-03 15:48:02.309905803 +0000 UTC m=+8251.513249230" Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.044312 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.046060 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.046213 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.047405 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.047569 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d" gracePeriod=600 Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.339393 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d" exitCode=0 Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.339759 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d"} Oct 03 15:48:06 crc kubenswrapper[4959]: I1003 15:48:06.339800 4959 scope.go:117] "RemoveContainer" containerID="c0f82bd1378bba78266b1de9031c3514e6d462265a207d0bc3288437ea5759f4" Oct 03 15:48:07 crc kubenswrapper[4959]: I1003 15:48:07.353684 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236"} Oct 03 15:48:08 crc kubenswrapper[4959]: I1003 15:48:08.008920 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:08 crc kubenswrapper[4959]: I1003 15:48:08.009410 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:08 crc kubenswrapper[4959]: I1003 15:48:08.070864 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:08 crc kubenswrapper[4959]: I1003 15:48:08.447824 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:08 crc kubenswrapper[4959]: I1003 15:48:08.498429 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:48:10 crc kubenswrapper[4959]: I1003 15:48:10.383570 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7lxxq" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="registry-server" containerID="cri-o://99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c" gracePeriod=2 Oct 03 15:48:10 crc kubenswrapper[4959]: I1003 15:48:10.929232 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.049366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content\") pod \"a348127e-b073-4ce9-b496-b36c0d704c6f\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.049745 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities\") pod \"a348127e-b073-4ce9-b496-b36c0d704c6f\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.049914 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl6dc\" (UniqueName: \"kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc\") pod \"a348127e-b073-4ce9-b496-b36c0d704c6f\" (UID: \"a348127e-b073-4ce9-b496-b36c0d704c6f\") " Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.050525 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities" (OuterVolumeSpecName: "utilities") pod "a348127e-b073-4ce9-b496-b36c0d704c6f" (UID: "a348127e-b073-4ce9-b496-b36c0d704c6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.050708 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.073306 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc" (OuterVolumeSpecName: "kube-api-access-zl6dc") pod "a348127e-b073-4ce9-b496-b36c0d704c6f" (UID: "a348127e-b073-4ce9-b496-b36c0d704c6f"). InnerVolumeSpecName "kube-api-access-zl6dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.078686 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a348127e-b073-4ce9-b496-b36c0d704c6f" (UID: "a348127e-b073-4ce9-b496-b36c0d704c6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.152716 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl6dc\" (UniqueName: \"kubernetes.io/projected/a348127e-b073-4ce9-b496-b36c0d704c6f-kube-api-access-zl6dc\") on node \"crc\" DevicePath \"\"" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.152754 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a348127e-b073-4ce9-b496-b36c0d704c6f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.393606 4959 generic.go:334] "Generic (PLEG): container finished" podID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerID="99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c" exitCode=0 Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.393654 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerDied","Data":"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c"} Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.393692 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lxxq" event={"ID":"a348127e-b073-4ce9-b496-b36c0d704c6f","Type":"ContainerDied","Data":"15fb442af3f8625769582423811958997d7a01679019f78fbadb43b302755264"} Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.393715 4959 scope.go:117] "RemoveContainer" containerID="99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.393798 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lxxq" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.413705 4959 scope.go:117] "RemoveContainer" containerID="473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.443327 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.452355 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lxxq"] Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.459198 4959 scope.go:117] "RemoveContainer" containerID="50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.487477 4959 scope.go:117] "RemoveContainer" containerID="99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c" Oct 03 15:48:11 crc kubenswrapper[4959]: E1003 15:48:11.487983 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c\": container with ID starting with 99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c not found: ID does not exist" containerID="99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.488033 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c"} err="failed to get container status \"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c\": rpc error: code = NotFound desc = could not find container \"99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c\": container with ID starting with 99fd23ce193ebfd26b09dab4459c5f5ab2597fff08a17f361c420670ce698a5c not found: ID does not exist" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.488062 4959 scope.go:117] "RemoveContainer" containerID="473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427" Oct 03 15:48:11 crc kubenswrapper[4959]: E1003 15:48:11.488515 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427\": container with ID starting with 473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427 not found: ID does not exist" containerID="473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.488564 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427"} err="failed to get container status \"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427\": rpc error: code = NotFound desc = could not find container \"473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427\": container with ID starting with 473d29019d36eaf2432103ab22b4a956d7d23c3a1bf23c0f59ca51fd9e6c8427 not found: ID does not exist" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.488592 4959 scope.go:117] "RemoveContainer" containerID="50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30" Oct 03 15:48:11 crc kubenswrapper[4959]: E1003 15:48:11.488949 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30\": container with ID starting with 50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30 not found: ID does not exist" containerID="50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.488973 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30"} err="failed to get container status \"50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30\": rpc error: code = NotFound desc = could not find container \"50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30\": container with ID starting with 50004a4195260d0b5c414d03056e004f6a797556199f2e7beca86c6cf85d9c30 not found: ID does not exist" Oct 03 15:48:11 crc kubenswrapper[4959]: I1003 15:48:11.703855 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" path="/var/lib/kubelet/pods/a348127e-b073-4ce9-b496-b36c0d704c6f/volumes" Oct 03 15:50:06 crc kubenswrapper[4959]: I1003 15:50:06.043931 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:50:06 crc kubenswrapper[4959]: I1003 15:50:06.045228 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:50:36 crc kubenswrapper[4959]: I1003 15:50:36.044311 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:50:36 crc kubenswrapper[4959]: I1003 15:50:36.045035 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.044375 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.045020 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.045078 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.046145 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.046246 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" gracePeriod=600 Oct 03 15:51:06 crc kubenswrapper[4959]: E1003 15:51:06.248882 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.382878 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" exitCode=0 Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.382932 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236"} Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.382970 4959 scope.go:117] "RemoveContainer" containerID="0c363b9fa657af2ff48b3a121848184368cdd3b215fb0e1092f83791ba54017d" Oct 03 15:51:06 crc kubenswrapper[4959]: I1003 15:51:06.384009 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:51:06 crc kubenswrapper[4959]: E1003 15:51:06.384808 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:51:17 crc kubenswrapper[4959]: I1003 15:51:17.687137 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:51:17 crc kubenswrapper[4959]: E1003 15:51:17.687696 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:51:29 crc kubenswrapper[4959]: I1003 15:51:29.686340 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:51:29 crc kubenswrapper[4959]: E1003 15:51:29.687517 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:51:39 crc kubenswrapper[4959]: I1003 15:51:39.806644 4959 generic.go:334] "Generic (PLEG): container finished" podID="2a09699c-f954-45ac-b96b-accef4753ab3" containerID="1425b6f659d747820a2da6575cd116fdcfbcc7774839c0db2fafcc1875abb4a1" exitCode=0 Oct 03 15:51:39 crc kubenswrapper[4959]: I1003 15:51:39.806753 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" event={"ID":"2a09699c-f954-45ac-b96b-accef4753ab3","Type":"ContainerDied","Data":"1425b6f659d747820a2da6575cd116fdcfbcc7774839c0db2fafcc1875abb4a1"} Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.321080 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454564 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454649 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454698 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454733 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454805 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454861 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454881 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbf6m\" (UniqueName: \"kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454945 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.454988 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.455019 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.455046 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory\") pod \"2a09699c-f954-45ac-b96b-accef4753ab3\" (UID: \"2a09699c-f954-45ac-b96b-accef4753ab3\") " Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.460010 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.460575 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m" (OuterVolumeSpecName: "kube-api-access-xbf6m") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "kube-api-access-xbf6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.461888 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph" (OuterVolumeSpecName: "ceph") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.484863 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.487407 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory" (OuterVolumeSpecName: "inventory") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.491149 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.494518 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.495252 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.502123 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.507704 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.510874 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2a09699c-f954-45ac-b96b-accef4753ab3" (UID: "2a09699c-f954-45ac-b96b-accef4753ab3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558665 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558716 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558730 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558741 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558752 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558763 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558773 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558782 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558794 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbf6m\" (UniqueName: \"kubernetes.io/projected/2a09699c-f954-45ac-b96b-accef4753ab3-kube-api-access-xbf6m\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558806 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.558816 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2a09699c-f954-45ac-b96b-accef4753ab3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.700622 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:51:41 crc kubenswrapper[4959]: E1003 15:51:41.700932 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.829788 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" event={"ID":"2a09699c-f954-45ac-b96b-accef4753ab3","Type":"ContainerDied","Data":"14acbd158741e1a5138328bff85654d54628c0b857b697177460d0c0511b8506"} Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.829833 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14acbd158741e1a5138328bff85654d54628c0b857b697177460d0c0511b8506" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.829893 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8w98r" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.931234 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-h4zpc"] Oct 03 15:51:41 crc kubenswrapper[4959]: E1003 15:51:41.931936 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="extract-content" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.931966 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="extract-content" Oct 03 15:51:41 crc kubenswrapper[4959]: E1003 15:51:41.932013 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="extract-utilities" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.932024 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="extract-utilities" Oct 03 15:51:41 crc kubenswrapper[4959]: E1003 15:51:41.932066 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="registry-server" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.932078 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="registry-server" Oct 03 15:51:41 crc kubenswrapper[4959]: E1003 15:51:41.932117 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a09699c-f954-45ac-b96b-accef4753ab3" containerName="nova-cell1-openstack-openstack-cell1" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.932129 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a09699c-f954-45ac-b96b-accef4753ab3" containerName="nova-cell1-openstack-openstack-cell1" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.932485 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a09699c-f954-45ac-b96b-accef4753ab3" containerName="nova-cell1-openstack-openstack-cell1" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.932520 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="a348127e-b073-4ce9-b496-b36c0d704c6f" containerName="registry-server" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.933708 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.940763 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.940815 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.940779 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.941159 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.942542 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:51:41 crc kubenswrapper[4959]: I1003 15:51:41.943751 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-h4zpc"] Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.069240 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.069312 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wq45\" (UniqueName: \"kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.069398 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.069765 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.069900 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.070013 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.070130 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.070275 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172355 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172418 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wq45\" (UniqueName: \"kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172457 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172503 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172539 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172632 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.172680 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.177501 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.178707 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.178971 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.179062 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.180605 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.181578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.185058 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.194621 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wq45\" (UniqueName: \"kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45\") pod \"telemetry-openstack-openstack-cell1-h4zpc\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.268819 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.850429 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-h4zpc"] Oct 03 15:51:42 crc kubenswrapper[4959]: I1003 15:51:42.854494 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:51:43 crc kubenswrapper[4959]: I1003 15:51:43.856908 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" event={"ID":"ca1f3663-0568-4600-9c69-8a92589523ca","Type":"ContainerStarted","Data":"f0519b88bff56f3cfa6d9f0de70f2b76f4e98888ea8fea1d5e44ffdbbab15e8d"} Oct 03 15:51:43 crc kubenswrapper[4959]: I1003 15:51:43.857256 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" event={"ID":"ca1f3663-0568-4600-9c69-8a92589523ca","Type":"ContainerStarted","Data":"5f082e12265895955f8b96faa4b2a7bb931300e474fb4aa87fdbf93144025426"} Oct 03 15:51:43 crc kubenswrapper[4959]: I1003 15:51:43.889835 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" podStartSLOduration=2.700313257 podStartE2EDuration="2.889808527s" podCreationTimestamp="2025-10-03 15:51:41 +0000 UTC" firstStartedPulling="2025-10-03 15:51:42.854213526 +0000 UTC m=+8472.057556953" lastFinishedPulling="2025-10-03 15:51:43.043708806 +0000 UTC m=+8472.247052223" observedRunningTime="2025-10-03 15:51:43.872263068 +0000 UTC m=+8473.075606495" watchObservedRunningTime="2025-10-03 15:51:43.889808527 +0000 UTC m=+8473.093151954" Oct 03 15:51:56 crc kubenswrapper[4959]: I1003 15:51:56.687228 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:51:56 crc kubenswrapper[4959]: E1003 15:51:56.689900 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:52:09 crc kubenswrapper[4959]: I1003 15:52:09.686067 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:52:09 crc kubenswrapper[4959]: E1003 15:52:09.686787 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:52:21 crc kubenswrapper[4959]: I1003 15:52:21.702312 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:52:21 crc kubenswrapper[4959]: E1003 15:52:21.703498 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:52:33 crc kubenswrapper[4959]: I1003 15:52:33.686119 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:52:33 crc kubenswrapper[4959]: E1003 15:52:33.687058 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:52:46 crc kubenswrapper[4959]: I1003 15:52:46.686571 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:52:46 crc kubenswrapper[4959]: E1003 15:52:46.687442 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:52:57 crc kubenswrapper[4959]: I1003 15:52:57.688072 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:52:57 crc kubenswrapper[4959]: E1003 15:52:57.689038 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:53:09 crc kubenswrapper[4959]: I1003 15:53:09.686975 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:53:09 crc kubenswrapper[4959]: E1003 15:53:09.687997 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:53:21 crc kubenswrapper[4959]: I1003 15:53:21.694692 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:53:21 crc kubenswrapper[4959]: E1003 15:53:21.698488 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:53:35 crc kubenswrapper[4959]: I1003 15:53:35.686009 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:53:35 crc kubenswrapper[4959]: E1003 15:53:35.686968 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:53:50 crc kubenswrapper[4959]: I1003 15:53:50.687478 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:53:50 crc kubenswrapper[4959]: E1003 15:53:50.688586 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.263217 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.266096 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.275442 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.406919 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9v55\" (UniqueName: \"kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.407004 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.407212 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.510344 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9v55\" (UniqueName: \"kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.510467 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.510538 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.511069 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.511230 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.530664 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9v55\" (UniqueName: \"kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55\") pod \"community-operators-r7dbx\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:56 crc kubenswrapper[4959]: I1003 15:53:56.593502 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:53:57 crc kubenswrapper[4959]: I1003 15:53:57.122392 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:53:57 crc kubenswrapper[4959]: I1003 15:53:57.492541 4959 generic.go:334] "Generic (PLEG): container finished" podID="3fe3162f-591b-461b-9837-7ce87acc2607" containerID="05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6" exitCode=0 Oct 03 15:53:57 crc kubenswrapper[4959]: I1003 15:53:57.492607 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerDied","Data":"05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6"} Oct 03 15:53:57 crc kubenswrapper[4959]: I1003 15:53:57.492642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerStarted","Data":"733e540cfa78bb6dcef3f46690367fcfbf4ddd6e22e593b4383366479a955feb"} Oct 03 15:53:58 crc kubenswrapper[4959]: I1003 15:53:58.509169 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerStarted","Data":"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565"} Oct 03 15:53:59 crc kubenswrapper[4959]: I1003 15:53:59.522869 4959 generic.go:334] "Generic (PLEG): container finished" podID="3fe3162f-591b-461b-9837-7ce87acc2607" containerID="cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565" exitCode=0 Oct 03 15:53:59 crc kubenswrapper[4959]: I1003 15:53:59.522959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerDied","Data":"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565"} Oct 03 15:54:00 crc kubenswrapper[4959]: I1003 15:54:00.539557 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerStarted","Data":"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65"} Oct 03 15:54:00 crc kubenswrapper[4959]: I1003 15:54:00.567765 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r7dbx" podStartSLOduration=2.173171297 podStartE2EDuration="4.567743839s" podCreationTimestamp="2025-10-03 15:53:56 +0000 UTC" firstStartedPulling="2025-10-03 15:53:57.500497793 +0000 UTC m=+8606.703841210" lastFinishedPulling="2025-10-03 15:53:59.895070335 +0000 UTC m=+8609.098413752" observedRunningTime="2025-10-03 15:54:00.563805843 +0000 UTC m=+8609.767149330" watchObservedRunningTime="2025-10-03 15:54:00.567743839 +0000 UTC m=+8609.771087256" Oct 03 15:54:01 crc kubenswrapper[4959]: I1003 15:54:01.686700 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:54:01 crc kubenswrapper[4959]: E1003 15:54:01.687477 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:54:06 crc kubenswrapper[4959]: I1003 15:54:06.594566 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:06 crc kubenswrapper[4959]: I1003 15:54:06.595234 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:06 crc kubenswrapper[4959]: I1003 15:54:06.644707 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:06 crc kubenswrapper[4959]: I1003 15:54:06.716680 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:06 crc kubenswrapper[4959]: I1003 15:54:06.888988 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:54:08 crc kubenswrapper[4959]: I1003 15:54:08.628025 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r7dbx" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="registry-server" containerID="cri-o://f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65" gracePeriod=2 Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.146988 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.321572 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content\") pod \"3fe3162f-591b-461b-9837-7ce87acc2607\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.321983 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9v55\" (UniqueName: \"kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55\") pod \"3fe3162f-591b-461b-9837-7ce87acc2607\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.322071 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities\") pod \"3fe3162f-591b-461b-9837-7ce87acc2607\" (UID: \"3fe3162f-591b-461b-9837-7ce87acc2607\") " Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.323142 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities" (OuterVolumeSpecName: "utilities") pod "3fe3162f-591b-461b-9837-7ce87acc2607" (UID: "3fe3162f-591b-461b-9837-7ce87acc2607"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.323800 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.333421 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55" (OuterVolumeSpecName: "kube-api-access-l9v55") pod "3fe3162f-591b-461b-9837-7ce87acc2607" (UID: "3fe3162f-591b-461b-9837-7ce87acc2607"). InnerVolumeSpecName "kube-api-access-l9v55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.425989 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9v55\" (UniqueName: \"kubernetes.io/projected/3fe3162f-591b-461b-9837-7ce87acc2607-kube-api-access-l9v55\") on node \"crc\" DevicePath \"\"" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.537864 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fe3162f-591b-461b-9837-7ce87acc2607" (UID: "3fe3162f-591b-461b-9837-7ce87acc2607"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.630027 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fe3162f-591b-461b-9837-7ce87acc2607-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.641538 4959 generic.go:334] "Generic (PLEG): container finished" podID="3fe3162f-591b-461b-9837-7ce87acc2607" containerID="f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65" exitCode=0 Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.641588 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerDied","Data":"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65"} Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.641613 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7dbx" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.641618 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7dbx" event={"ID":"3fe3162f-591b-461b-9837-7ce87acc2607","Type":"ContainerDied","Data":"733e540cfa78bb6dcef3f46690367fcfbf4ddd6e22e593b4383366479a955feb"} Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.641643 4959 scope.go:117] "RemoveContainer" containerID="f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.678552 4959 scope.go:117] "RemoveContainer" containerID="cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.679560 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.700352 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r7dbx"] Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.718849 4959 scope.go:117] "RemoveContainer" containerID="05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.760546 4959 scope.go:117] "RemoveContainer" containerID="f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65" Oct 03 15:54:09 crc kubenswrapper[4959]: E1003 15:54:09.761107 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65\": container with ID starting with f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65 not found: ID does not exist" containerID="f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.761164 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65"} err="failed to get container status \"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65\": rpc error: code = NotFound desc = could not find container \"f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65\": container with ID starting with f52cd5e9c0af63802944883912cb1cbcab05204ea8a43f97fec955a1c7cacb65 not found: ID does not exist" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.761246 4959 scope.go:117] "RemoveContainer" containerID="cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565" Oct 03 15:54:09 crc kubenswrapper[4959]: E1003 15:54:09.761695 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565\": container with ID starting with cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565 not found: ID does not exist" containerID="cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.761738 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565"} err="failed to get container status \"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565\": rpc error: code = NotFound desc = could not find container \"cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565\": container with ID starting with cc9f4fbec3bb63e85ea670ad9a5f4dba47d79baef5b749521cc0d8e074a1b565 not found: ID does not exist" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.761767 4959 scope.go:117] "RemoveContainer" containerID="05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6" Oct 03 15:54:09 crc kubenswrapper[4959]: E1003 15:54:09.762357 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6\": container with ID starting with 05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6 not found: ID does not exist" containerID="05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6" Oct 03 15:54:09 crc kubenswrapper[4959]: I1003 15:54:09.762400 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6"} err="failed to get container status \"05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6\": rpc error: code = NotFound desc = could not find container \"05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6\": container with ID starting with 05b1a4287ea20c045022b079fb93f06ed1a53f6b9ce234fcf0e5f799039bfcd6 not found: ID does not exist" Oct 03 15:54:11 crc kubenswrapper[4959]: I1003 15:54:11.710373 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" path="/var/lib/kubelet/pods/3fe3162f-591b-461b-9837-7ce87acc2607/volumes" Oct 03 15:54:12 crc kubenswrapper[4959]: I1003 15:54:12.685736 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:54:12 crc kubenswrapper[4959]: E1003 15:54:12.685944 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:54:27 crc kubenswrapper[4959]: I1003 15:54:27.685347 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:54:27 crc kubenswrapper[4959]: E1003 15:54:27.686043 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:54:40 crc kubenswrapper[4959]: I1003 15:54:40.686788 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:54:40 crc kubenswrapper[4959]: E1003 15:54:40.687871 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:54:54 crc kubenswrapper[4959]: I1003 15:54:54.687344 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:54:54 crc kubenswrapper[4959]: E1003 15:54:54.690863 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:55:07 crc kubenswrapper[4959]: I1003 15:55:07.687662 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:55:07 crc kubenswrapper[4959]: E1003 15:55:07.688781 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:55:18 crc kubenswrapper[4959]: I1003 15:55:18.686032 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:55:18 crc kubenswrapper[4959]: E1003 15:55:18.687420 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.062846 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:23 crc kubenswrapper[4959]: E1003 15:55:23.063624 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="extract-content" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.063638 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="extract-content" Oct 03 15:55:23 crc kubenswrapper[4959]: E1003 15:55:23.063651 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="extract-utilities" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.063659 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="extract-utilities" Oct 03 15:55:23 crc kubenswrapper[4959]: E1003 15:55:23.063682 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="registry-server" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.063689 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="registry-server" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.063960 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe3162f-591b-461b-9837-7ce87acc2607" containerName="registry-server" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.065865 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.080062 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.251585 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmqgs\" (UniqueName: \"kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.251691 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.251789 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.354106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.354277 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.354492 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmqgs\" (UniqueName: \"kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.354716 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.354910 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.373875 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmqgs\" (UniqueName: \"kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs\") pod \"redhat-operators-4j2j9\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.402709 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:23 crc kubenswrapper[4959]: I1003 15:55:23.915373 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:24 crc kubenswrapper[4959]: I1003 15:55:24.587768 4959 generic.go:334] "Generic (PLEG): container finished" podID="f80389d2-791a-4bc9-a4c3-369be122a179" containerID="38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1" exitCode=0 Oct 03 15:55:24 crc kubenswrapper[4959]: I1003 15:55:24.588175 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerDied","Data":"38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1"} Oct 03 15:55:24 crc kubenswrapper[4959]: I1003 15:55:24.588269 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerStarted","Data":"59f778711b2818bd2cfd5f3c18352664e59b0bf0126ff4e84aea40c00a76108d"} Oct 03 15:55:26 crc kubenswrapper[4959]: I1003 15:55:26.607172 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerStarted","Data":"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6"} Oct 03 15:55:27 crc kubenswrapper[4959]: I1003 15:55:27.619582 4959 generic.go:334] "Generic (PLEG): container finished" podID="f80389d2-791a-4bc9-a4c3-369be122a179" containerID="bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6" exitCode=0 Oct 03 15:55:27 crc kubenswrapper[4959]: I1003 15:55:27.619673 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerDied","Data":"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6"} Oct 03 15:55:29 crc kubenswrapper[4959]: I1003 15:55:29.650971 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerStarted","Data":"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f"} Oct 03 15:55:29 crc kubenswrapper[4959]: I1003 15:55:29.694430 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4j2j9" podStartSLOduration=2.549943871 podStartE2EDuration="6.694407815s" podCreationTimestamp="2025-10-03 15:55:23 +0000 UTC" firstStartedPulling="2025-10-03 15:55:24.592618073 +0000 UTC m=+8693.795961520" lastFinishedPulling="2025-10-03 15:55:28.737082047 +0000 UTC m=+8697.940425464" observedRunningTime="2025-10-03 15:55:29.675266727 +0000 UTC m=+8698.878610184" watchObservedRunningTime="2025-10-03 15:55:29.694407815 +0000 UTC m=+8698.897751252" Oct 03 15:55:33 crc kubenswrapper[4959]: I1003 15:55:33.403022 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:33 crc kubenswrapper[4959]: I1003 15:55:33.403517 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:33 crc kubenswrapper[4959]: I1003 15:55:33.685734 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:55:33 crc kubenswrapper[4959]: E1003 15:55:33.686897 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:55:34 crc kubenswrapper[4959]: I1003 15:55:34.475971 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4j2j9" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="registry-server" probeResult="failure" output=< Oct 03 15:55:34 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 15:55:34 crc kubenswrapper[4959]: > Oct 03 15:55:43 crc kubenswrapper[4959]: I1003 15:55:43.488497 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:43 crc kubenswrapper[4959]: I1003 15:55:43.586666 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:43 crc kubenswrapper[4959]: I1003 15:55:43.745571 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:44 crc kubenswrapper[4959]: I1003 15:55:44.838388 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4j2j9" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="registry-server" containerID="cri-o://d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f" gracePeriod=2 Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.688088 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:55:45 crc kubenswrapper[4959]: E1003 15:55:45.689435 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.706945 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.802940 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities\") pod \"f80389d2-791a-4bc9-a4c3-369be122a179\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.803048 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content\") pod \"f80389d2-791a-4bc9-a4c3-369be122a179\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.803575 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmqgs\" (UniqueName: \"kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs\") pod \"f80389d2-791a-4bc9-a4c3-369be122a179\" (UID: \"f80389d2-791a-4bc9-a4c3-369be122a179\") " Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.803862 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities" (OuterVolumeSpecName: "utilities") pod "f80389d2-791a-4bc9-a4c3-369be122a179" (UID: "f80389d2-791a-4bc9-a4c3-369be122a179"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.805372 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.812430 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs" (OuterVolumeSpecName: "kube-api-access-nmqgs") pod "f80389d2-791a-4bc9-a4c3-369be122a179" (UID: "f80389d2-791a-4bc9-a4c3-369be122a179"). InnerVolumeSpecName "kube-api-access-nmqgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.853728 4959 generic.go:334] "Generic (PLEG): container finished" podID="f80389d2-791a-4bc9-a4c3-369be122a179" containerID="d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f" exitCode=0 Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.853776 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerDied","Data":"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f"} Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.853823 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2j9" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.853852 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2j9" event={"ID":"f80389d2-791a-4bc9-a4c3-369be122a179","Type":"ContainerDied","Data":"59f778711b2818bd2cfd5f3c18352664e59b0bf0126ff4e84aea40c00a76108d"} Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.853880 4959 scope.go:117] "RemoveContainer" containerID="d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.902513 4959 scope.go:117] "RemoveContainer" containerID="bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.907787 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmqgs\" (UniqueName: \"kubernetes.io/projected/f80389d2-791a-4bc9-a4c3-369be122a179-kube-api-access-nmqgs\") on node \"crc\" DevicePath \"\"" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.928462 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f80389d2-791a-4bc9-a4c3-369be122a179" (UID: "f80389d2-791a-4bc9-a4c3-369be122a179"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:55:45 crc kubenswrapper[4959]: I1003 15:55:45.953463 4959 scope.go:117] "RemoveContainer" containerID="38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:45.999243 4959 scope.go:117] "RemoveContainer" containerID="d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f" Oct 03 15:55:46 crc kubenswrapper[4959]: E1003 15:55:45.999735 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f\": container with ID starting with d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f not found: ID does not exist" containerID="d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:45.999760 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f"} err="failed to get container status \"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f\": rpc error: code = NotFound desc = could not find container \"d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f\": container with ID starting with d252dea0f9e5b24d6130b07f2fd0d7729a3bf6b8cd6983f924c7112a27a0bf1f not found: ID does not exist" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:45.999779 4959 scope.go:117] "RemoveContainer" containerID="bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6" Oct 03 15:55:46 crc kubenswrapper[4959]: E1003 15:55:46.000036 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6\": container with ID starting with bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6 not found: ID does not exist" containerID="bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.000053 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6"} err="failed to get container status \"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6\": rpc error: code = NotFound desc = could not find container \"bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6\": container with ID starting with bc1d60a750b5769d063e0cfd5ac6fb0e1b735421f9dfcab4814e95196b089af6 not found: ID does not exist" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.000066 4959 scope.go:117] "RemoveContainer" containerID="38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1" Oct 03 15:55:46 crc kubenswrapper[4959]: E1003 15:55:46.000327 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1\": container with ID starting with 38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1 not found: ID does not exist" containerID="38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.000354 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1"} err="failed to get container status \"38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1\": rpc error: code = NotFound desc = could not find container \"38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1\": container with ID starting with 38b6f8ac14b283cf4005a16448d1467f55795842d90223172e49fbe8901396b1 not found: ID does not exist" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.009917 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f80389d2-791a-4bc9-a4c3-369be122a179-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.196460 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:46 crc kubenswrapper[4959]: I1003 15:55:46.206263 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4j2j9"] Oct 03 15:55:47 crc kubenswrapper[4959]: I1003 15:55:47.702605 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" path="/var/lib/kubelet/pods/f80389d2-791a-4bc9-a4c3-369be122a179/volumes" Oct 03 15:56:00 crc kubenswrapper[4959]: I1003 15:56:00.686382 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:56:00 crc kubenswrapper[4959]: E1003 15:56:00.687402 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 15:56:02 crc kubenswrapper[4959]: I1003 15:56:02.061462 4959 generic.go:334] "Generic (PLEG): container finished" podID="ca1f3663-0568-4600-9c69-8a92589523ca" containerID="f0519b88bff56f3cfa6d9f0de70f2b76f4e98888ea8fea1d5e44ffdbbab15e8d" exitCode=0 Oct 03 15:56:02 crc kubenswrapper[4959]: I1003 15:56:02.061534 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" event={"ID":"ca1f3663-0568-4600-9c69-8a92589523ca","Type":"ContainerDied","Data":"f0519b88bff56f3cfa6d9f0de70f2b76f4e98888ea8fea1d5e44ffdbbab15e8d"} Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.529504 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611664 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611707 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611859 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wq45\" (UniqueName: \"kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611890 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611948 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.611971 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.612312 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.613063 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2\") pod \"ca1f3663-0568-4600-9c69-8a92589523ca\" (UID: \"ca1f3663-0568-4600-9c69-8a92589523ca\") " Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.617822 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph" (OuterVolumeSpecName: "ceph") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.617970 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.618425 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45" (OuterVolumeSpecName: "kube-api-access-2wq45") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "kube-api-access-2wq45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.643287 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.644680 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.645567 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.645876 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory" (OuterVolumeSpecName: "inventory") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.650717 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ca1f3663-0568-4600-9c69-8a92589523ca" (UID: "ca1f3663-0568-4600-9c69-8a92589523ca"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716173 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716222 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716232 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716243 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wq45\" (UniqueName: \"kubernetes.io/projected/ca1f3663-0568-4600-9c69-8a92589523ca-kube-api-access-2wq45\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716254 4959 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716263 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716273 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:03 crc kubenswrapper[4959]: I1003 15:56:03.716281 4959 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca1f3663-0568-4600-9c69-8a92589523ca-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.084728 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" event={"ID":"ca1f3663-0568-4600-9c69-8a92589523ca","Type":"ContainerDied","Data":"5f082e12265895955f8b96faa4b2a7bb931300e474fb4aa87fdbf93144025426"} Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.084774 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f082e12265895955f8b96faa4b2a7bb931300e474fb4aa87fdbf93144025426" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.084810 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-h4zpc" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.203050 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-chj9r"] Oct 03 15:56:04 crc kubenswrapper[4959]: E1003 15:56:04.203606 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="registry-server" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.203629 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="registry-server" Oct 03 15:56:04 crc kubenswrapper[4959]: E1003 15:56:04.203652 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="extract-content" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.203663 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="extract-content" Oct 03 15:56:04 crc kubenswrapper[4959]: E1003 15:56:04.203683 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="extract-utilities" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.203691 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="extract-utilities" Oct 03 15:56:04 crc kubenswrapper[4959]: E1003 15:56:04.203722 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca1f3663-0568-4600-9c69-8a92589523ca" containerName="telemetry-openstack-openstack-cell1" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.203730 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca1f3663-0568-4600-9c69-8a92589523ca" containerName="telemetry-openstack-openstack-cell1" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.204143 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80389d2-791a-4bc9-a4c3-369be122a179" containerName="registry-server" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.204178 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca1f3663-0568-4600-9c69-8a92589523ca" containerName="telemetry-openstack-openstack-cell1" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.205131 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.207834 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.208090 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.208414 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.208619 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.209882 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.216956 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-chj9r"] Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329128 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329232 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54m45\" (UniqueName: \"kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329308 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329401 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329480 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.329599 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.430878 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.430927 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54m45\" (UniqueName: \"kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.430966 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.431024 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.431073 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.431146 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.437264 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.437299 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.437341 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.438317 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.439559 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.450016 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54m45\" (UniqueName: \"kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45\") pod \"neutron-sriov-openstack-openstack-cell1-chj9r\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:04 crc kubenswrapper[4959]: I1003 15:56:04.521277 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:56:05 crc kubenswrapper[4959]: I1003 15:56:05.055874 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-chj9r"] Oct 03 15:56:05 crc kubenswrapper[4959]: I1003 15:56:05.096390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" event={"ID":"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7","Type":"ContainerStarted","Data":"a6243f3d7465badfa1f28312d71651e6c7292b96c391f2287d77852085f644c2"} Oct 03 15:56:06 crc kubenswrapper[4959]: I1003 15:56:06.120591 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" event={"ID":"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7","Type":"ContainerStarted","Data":"4e099e7342a3341dfd4f030f60483306bc84f939c89ab04a335719dcb158ab02"} Oct 03 15:56:06 crc kubenswrapper[4959]: I1003 15:56:06.142508 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" podStartSLOduration=1.937157151 podStartE2EDuration="2.142486328s" podCreationTimestamp="2025-10-03 15:56:04 +0000 UTC" firstStartedPulling="2025-10-03 15:56:05.071593855 +0000 UTC m=+8734.274937272" lastFinishedPulling="2025-10-03 15:56:05.276923032 +0000 UTC m=+8734.480266449" observedRunningTime="2025-10-03 15:56:06.139240439 +0000 UTC m=+8735.342583856" watchObservedRunningTime="2025-10-03 15:56:06.142486328 +0000 UTC m=+8735.345829755" Oct 03 15:56:12 crc kubenswrapper[4959]: I1003 15:56:12.686473 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:56:13 crc kubenswrapper[4959]: I1003 15:56:13.203440 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78"} Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.518741 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.522668 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.538972 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.700671 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.701208 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.701393 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr6zj\" (UniqueName: \"kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.803826 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.803917 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr6zj\" (UniqueName: \"kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.804293 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.804708 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.805405 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.832224 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr6zj\" (UniqueName: \"kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj\") pod \"certified-operators-p47z4\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:51 crc kubenswrapper[4959]: I1003 15:57:51.860540 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:57:52 crc kubenswrapper[4959]: I1003 15:57:52.197654 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:57:52 crc kubenswrapper[4959]: W1003 15:57:52.202295 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc13528c1_b6ce_42de_b164_e1aa69f0d1df.slice/crio-5d7050e575897de7e49d153bcb1ea2e6cf4c296771562e9cb6cf8cda6762d99d WatchSource:0}: Error finding container 5d7050e575897de7e49d153bcb1ea2e6cf4c296771562e9cb6cf8cda6762d99d: Status 404 returned error can't find the container with id 5d7050e575897de7e49d153bcb1ea2e6cf4c296771562e9cb6cf8cda6762d99d Oct 03 15:57:52 crc kubenswrapper[4959]: I1003 15:57:52.526361 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerStarted","Data":"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54"} Oct 03 15:57:52 crc kubenswrapper[4959]: I1003 15:57:52.526410 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerStarted","Data":"5d7050e575897de7e49d153bcb1ea2e6cf4c296771562e9cb6cf8cda6762d99d"} Oct 03 15:57:52 crc kubenswrapper[4959]: E1003 15:57:52.628665 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc13528c1_b6ce_42de_b164_e1aa69f0d1df.slice/crio-0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54.scope\": RecentStats: unable to find data in memory cache]" Oct 03 15:57:53 crc kubenswrapper[4959]: I1003 15:57:53.543281 4959 generic.go:334] "Generic (PLEG): container finished" podID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerID="0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54" exitCode=0 Oct 03 15:57:53 crc kubenswrapper[4959]: I1003 15:57:53.543382 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerDied","Data":"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54"} Oct 03 15:57:53 crc kubenswrapper[4959]: I1003 15:57:53.546521 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 15:57:55 crc kubenswrapper[4959]: I1003 15:57:55.567136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerStarted","Data":"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403"} Oct 03 15:57:56 crc kubenswrapper[4959]: I1003 15:57:56.586684 4959 generic.go:334] "Generic (PLEG): container finished" podID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerID="14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403" exitCode=0 Oct 03 15:57:56 crc kubenswrapper[4959]: I1003 15:57:56.586748 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerDied","Data":"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403"} Oct 03 15:57:57 crc kubenswrapper[4959]: I1003 15:57:57.607888 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerStarted","Data":"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41"} Oct 03 15:57:57 crc kubenswrapper[4959]: I1003 15:57:57.628114 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p47z4" podStartSLOduration=2.837095037 podStartE2EDuration="6.628097325s" podCreationTimestamp="2025-10-03 15:57:51 +0000 UTC" firstStartedPulling="2025-10-03 15:57:53.546156309 +0000 UTC m=+8842.749499726" lastFinishedPulling="2025-10-03 15:57:57.337158587 +0000 UTC m=+8846.540502014" observedRunningTime="2025-10-03 15:57:57.626721151 +0000 UTC m=+8846.830064598" watchObservedRunningTime="2025-10-03 15:57:57.628097325 +0000 UTC m=+8846.831440742" Oct 03 15:58:01 crc kubenswrapper[4959]: I1003 15:58:01.860687 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:01 crc kubenswrapper[4959]: I1003 15:58:01.861565 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:01 crc kubenswrapper[4959]: I1003 15:58:01.937224 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:02 crc kubenswrapper[4959]: I1003 15:58:02.731168 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:02 crc kubenswrapper[4959]: I1003 15:58:02.789943 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:58:04 crc kubenswrapper[4959]: I1003 15:58:04.691418 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p47z4" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="registry-server" containerID="cri-o://a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41" gracePeriod=2 Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.196127 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.338386 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities\") pod \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.338532 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr6zj\" (UniqueName: \"kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj\") pod \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.338681 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content\") pod \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\" (UID: \"c13528c1-b6ce-42de-b164-e1aa69f0d1df\") " Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.339322 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities" (OuterVolumeSpecName: "utilities") pod "c13528c1-b6ce-42de-b164-e1aa69f0d1df" (UID: "c13528c1-b6ce-42de-b164-e1aa69f0d1df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.345896 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj" (OuterVolumeSpecName: "kube-api-access-zr6zj") pod "c13528c1-b6ce-42de-b164-e1aa69f0d1df" (UID: "c13528c1-b6ce-42de-b164-e1aa69f0d1df"). InnerVolumeSpecName "kube-api-access-zr6zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.390800 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c13528c1-b6ce-42de-b164-e1aa69f0d1df" (UID: "c13528c1-b6ce-42de-b164-e1aa69f0d1df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.441342 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr6zj\" (UniqueName: \"kubernetes.io/projected/c13528c1-b6ce-42de-b164-e1aa69f0d1df-kube-api-access-zr6zj\") on node \"crc\" DevicePath \"\"" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.441605 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.441619 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13528c1-b6ce-42de-b164-e1aa69f0d1df-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.708224 4959 generic.go:334] "Generic (PLEG): container finished" podID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerID="a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41" exitCode=0 Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.708301 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerDied","Data":"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41"} Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.708345 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p47z4" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.708390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p47z4" event={"ID":"c13528c1-b6ce-42de-b164-e1aa69f0d1df","Type":"ContainerDied","Data":"5d7050e575897de7e49d153bcb1ea2e6cf4c296771562e9cb6cf8cda6762d99d"} Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.708429 4959 scope.go:117] "RemoveContainer" containerID="a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.741965 4959 scope.go:117] "RemoveContainer" containerID="14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.768987 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.787555 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p47z4"] Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.790882 4959 scope.go:117] "RemoveContainer" containerID="0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.846096 4959 scope.go:117] "RemoveContainer" containerID="a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41" Oct 03 15:58:05 crc kubenswrapper[4959]: E1003 15:58:05.846678 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41\": container with ID starting with a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41 not found: ID does not exist" containerID="a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.846712 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41"} err="failed to get container status \"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41\": rpc error: code = NotFound desc = could not find container \"a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41\": container with ID starting with a0f1cd62e8909fa763874cf2afcbb6ac027f36faaaebe38ff56f0e5fc1b8bc41 not found: ID does not exist" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.846736 4959 scope.go:117] "RemoveContainer" containerID="14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403" Oct 03 15:58:05 crc kubenswrapper[4959]: E1003 15:58:05.847061 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403\": container with ID starting with 14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403 not found: ID does not exist" containerID="14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.847085 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403"} err="failed to get container status \"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403\": rpc error: code = NotFound desc = could not find container \"14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403\": container with ID starting with 14691b7b78bf560b51fb1c92216f4578a63bc72283489b87d9b4c29f14b6b403 not found: ID does not exist" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.847100 4959 scope.go:117] "RemoveContainer" containerID="0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54" Oct 03 15:58:05 crc kubenswrapper[4959]: E1003 15:58:05.847417 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54\": container with ID starting with 0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54 not found: ID does not exist" containerID="0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54" Oct 03 15:58:05 crc kubenswrapper[4959]: I1003 15:58:05.847446 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54"} err="failed to get container status \"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54\": rpc error: code = NotFound desc = could not find container \"0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54\": container with ID starting with 0634d68845390b348696507431f79ad9af7d997d5a83fe13e12ab7416b9f3b54 not found: ID does not exist" Oct 03 15:58:07 crc kubenswrapper[4959]: I1003 15:58:07.704639 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" path="/var/lib/kubelet/pods/c13528c1-b6ce-42de-b164-e1aa69f0d1df/volumes" Oct 03 15:58:36 crc kubenswrapper[4959]: I1003 15:58:36.044131 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:58:36 crc kubenswrapper[4959]: I1003 15:58:36.044899 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.493219 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:58:51 crc kubenswrapper[4959]: E1003 15:58:51.494335 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="registry-server" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.494349 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="registry-server" Oct 03 15:58:51 crc kubenswrapper[4959]: E1003 15:58:51.494366 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="extract-content" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.494373 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="extract-content" Oct 03 15:58:51 crc kubenswrapper[4959]: E1003 15:58:51.494406 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="extract-utilities" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.494413 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="extract-utilities" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.494633 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13528c1-b6ce-42de-b164-e1aa69f0d1df" containerName="registry-server" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.496500 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.524107 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.576357 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.576567 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn2cm\" (UniqueName: \"kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.576666 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.679030 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn2cm\" (UniqueName: \"kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.679154 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.679260 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.680077 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.680143 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.704158 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn2cm\" (UniqueName: \"kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm\") pod \"redhat-marketplace-lw8fl\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:51 crc kubenswrapper[4959]: I1003 15:58:51.833993 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:58:52 crc kubenswrapper[4959]: I1003 15:58:52.323656 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:58:53 crc kubenswrapper[4959]: I1003 15:58:53.291647 4959 generic.go:334] "Generic (PLEG): container finished" podID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerID="b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd" exitCode=0 Oct 03 15:58:53 crc kubenswrapper[4959]: I1003 15:58:53.291725 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerDied","Data":"b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd"} Oct 03 15:58:53 crc kubenswrapper[4959]: I1003 15:58:53.292136 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerStarted","Data":"1961f72cc17b1a66da858abe1deca89c94efb2c8cce62975a84a4245d395bfe4"} Oct 03 15:58:55 crc kubenswrapper[4959]: I1003 15:58:55.321991 4959 generic.go:334] "Generic (PLEG): container finished" podID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerID="fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4" exitCode=0 Oct 03 15:58:55 crc kubenswrapper[4959]: I1003 15:58:55.322069 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerDied","Data":"fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4"} Oct 03 15:58:56 crc kubenswrapper[4959]: I1003 15:58:56.343321 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerStarted","Data":"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6"} Oct 03 15:58:56 crc kubenswrapper[4959]: I1003 15:58:56.369795 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lw8fl" podStartSLOduration=2.938018253 podStartE2EDuration="5.369776984s" podCreationTimestamp="2025-10-03 15:58:51 +0000 UTC" firstStartedPulling="2025-10-03 15:58:53.294971313 +0000 UTC m=+8902.498314730" lastFinishedPulling="2025-10-03 15:58:55.726730044 +0000 UTC m=+8904.930073461" observedRunningTime="2025-10-03 15:58:56.359490833 +0000 UTC m=+8905.562834250" watchObservedRunningTime="2025-10-03 15:58:56.369776984 +0000 UTC m=+8905.573120401" Oct 03 15:59:01 crc kubenswrapper[4959]: I1003 15:59:01.834367 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:01 crc kubenswrapper[4959]: I1003 15:59:01.834944 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:01 crc kubenswrapper[4959]: I1003 15:59:01.902619 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:02 crc kubenswrapper[4959]: I1003 15:59:02.493321 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:02 crc kubenswrapper[4959]: I1003 15:59:02.558063 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:59:04 crc kubenswrapper[4959]: I1003 15:59:04.432755 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lw8fl" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="registry-server" containerID="cri-o://dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6" gracePeriod=2 Oct 03 15:59:04 crc kubenswrapper[4959]: I1003 15:59:04.968090 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.112998 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn2cm\" (UniqueName: \"kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm\") pod \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.113114 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content\") pod \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.113305 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities\") pod \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\" (UID: \"2aea0e32-3b7c-4b8e-be0c-6e89695d4676\") " Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.114550 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities" (OuterVolumeSpecName: "utilities") pod "2aea0e32-3b7c-4b8e-be0c-6e89695d4676" (UID: "2aea0e32-3b7c-4b8e-be0c-6e89695d4676"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.120121 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm" (OuterVolumeSpecName: "kube-api-access-bn2cm") pod "2aea0e32-3b7c-4b8e-be0c-6e89695d4676" (UID: "2aea0e32-3b7c-4b8e-be0c-6e89695d4676"). InnerVolumeSpecName "kube-api-access-bn2cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.133898 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2aea0e32-3b7c-4b8e-be0c-6e89695d4676" (UID: "2aea0e32-3b7c-4b8e-be0c-6e89695d4676"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.215291 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.215326 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn2cm\" (UniqueName: \"kubernetes.io/projected/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-kube-api-access-bn2cm\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.215336 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aea0e32-3b7c-4b8e-be0c-6e89695d4676-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.448058 4959 generic.go:334] "Generic (PLEG): container finished" podID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerID="dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6" exitCode=0 Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.448105 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerDied","Data":"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6"} Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.448129 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw8fl" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.448147 4959 scope.go:117] "RemoveContainer" containerID="dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.448135 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw8fl" event={"ID":"2aea0e32-3b7c-4b8e-be0c-6e89695d4676","Type":"ContainerDied","Data":"1961f72cc17b1a66da858abe1deca89c94efb2c8cce62975a84a4245d395bfe4"} Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.484703 4959 scope.go:117] "RemoveContainer" containerID="fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.491677 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.522754 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw8fl"] Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.531617 4959 scope.go:117] "RemoveContainer" containerID="b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.568551 4959 scope.go:117] "RemoveContainer" containerID="dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6" Oct 03 15:59:05 crc kubenswrapper[4959]: E1003 15:59:05.569161 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6\": container with ID starting with dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6 not found: ID does not exist" containerID="dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.569275 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6"} err="failed to get container status \"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6\": rpc error: code = NotFound desc = could not find container \"dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6\": container with ID starting with dc7f6f398834f3995a5d533cd3d26aedf60911b6d9ad5154d7cb4b15c3de01e6 not found: ID does not exist" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.569313 4959 scope.go:117] "RemoveContainer" containerID="fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4" Oct 03 15:59:05 crc kubenswrapper[4959]: E1003 15:59:05.569615 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4\": container with ID starting with fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4 not found: ID does not exist" containerID="fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.569737 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4"} err="failed to get container status \"fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4\": rpc error: code = NotFound desc = could not find container \"fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4\": container with ID starting with fba5c536885cc6b5d204416109ef4c28aa5e0b74d1efab0727e8a7eacc6cf9a4 not found: ID does not exist" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.569821 4959 scope.go:117] "RemoveContainer" containerID="b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd" Oct 03 15:59:05 crc kubenswrapper[4959]: E1003 15:59:05.570148 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd\": container with ID starting with b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd not found: ID does not exist" containerID="b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.570178 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd"} err="failed to get container status \"b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd\": rpc error: code = NotFound desc = could not find container \"b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd\": container with ID starting with b692f1621c7b8814c890cf325118155d5af408410ff3320c6c5b8059f2efcfcd not found: ID does not exist" Oct 03 15:59:05 crc kubenswrapper[4959]: I1003 15:59:05.703235 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" path="/var/lib/kubelet/pods/2aea0e32-3b7c-4b8e-be0c-6e89695d4676/volumes" Oct 03 15:59:06 crc kubenswrapper[4959]: I1003 15:59:06.044680 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:59:06 crc kubenswrapper[4959]: I1003 15:59:06.044774 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:59:24 crc kubenswrapper[4959]: I1003 15:59:24.678830 4959 generic.go:334] "Generic (PLEG): container finished" podID="fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" containerID="4e099e7342a3341dfd4f030f60483306bc84f939c89ab04a335719dcb158ab02" exitCode=0 Oct 03 15:59:24 crc kubenswrapper[4959]: I1003 15:59:24.678959 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" event={"ID":"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7","Type":"ContainerDied","Data":"4e099e7342a3341dfd4f030f60483306bc84f939c89ab04a335719dcb158ab02"} Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.185520 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.305943 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.306182 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54m45\" (UniqueName: \"kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.306416 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.306484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.306545 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.306715 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0\") pod \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\" (UID: \"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7\") " Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.313788 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45" (OuterVolumeSpecName: "kube-api-access-54m45") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "kube-api-access-54m45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.314520 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph" (OuterVolumeSpecName: "ceph") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.336307 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.352402 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.355293 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.356404 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory" (OuterVolumeSpecName: "inventory") pod "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" (UID: "fbd008c2-83fd-4c30-a3e6-6eb0e52965f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410113 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410158 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410175 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54m45\" (UniqueName: \"kubernetes.io/projected/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-kube-api-access-54m45\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410205 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410218 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.410229 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbd008c2-83fd-4c30-a3e6-6eb0e52965f7-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.704809 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" event={"ID":"fbd008c2-83fd-4c30-a3e6-6eb0e52965f7","Type":"ContainerDied","Data":"a6243f3d7465badfa1f28312d71651e6c7292b96c391f2287d77852085f644c2"} Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.704856 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6243f3d7465badfa1f28312d71651e6c7292b96c391f2287d77852085f644c2" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.704927 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-chj9r" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.807701 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq"] Oct 03 15:59:26 crc kubenswrapper[4959]: E1003 15:59:26.808378 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="extract-utilities" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808399 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="extract-utilities" Oct 03 15:59:26 crc kubenswrapper[4959]: E1003 15:59:26.808417 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="extract-content" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808425 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="extract-content" Oct 03 15:59:26 crc kubenswrapper[4959]: E1003 15:59:26.808466 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="registry-server" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808474 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="registry-server" Oct 03 15:59:26 crc kubenswrapper[4959]: E1003 15:59:26.808486 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808495 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808767 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbd008c2-83fd-4c30-a3e6-6eb0e52965f7" containerName="neutron-sriov-openstack-openstack-cell1" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.808797 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aea0e32-3b7c-4b8e-be0c-6e89695d4676" containerName="registry-server" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.810450 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.816217 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.816379 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.816406 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.816547 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.816659 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.823978 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq"] Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926080 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926161 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926258 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926288 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926537 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:26 crc kubenswrapper[4959]: I1003 15:59:26.926625 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rvhx\" (UniqueName: \"kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029387 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029452 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rvhx\" (UniqueName: \"kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029616 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029658 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029704 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.029738 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.034723 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.034886 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.034917 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.035363 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.037426 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.048480 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rvhx\" (UniqueName: \"kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx\") pod \"neutron-dhcp-openstack-openstack-cell1-2c2rq\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.128643 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 15:59:27 crc kubenswrapper[4959]: I1003 15:59:27.745814 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq"] Oct 03 15:59:28 crc kubenswrapper[4959]: I1003 15:59:28.740951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" event={"ID":"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4","Type":"ContainerStarted","Data":"287bfe353c16a7a1ea2e5a3d4939b0186094e3512f025c5cda2b9ca6e498bd21"} Oct 03 15:59:28 crc kubenswrapper[4959]: I1003 15:59:28.741717 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" event={"ID":"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4","Type":"ContainerStarted","Data":"c2f6ef76549236cf9d4ad94c8053de2ebb37528670e15267897bc349fb8679a0"} Oct 03 15:59:28 crc kubenswrapper[4959]: I1003 15:59:28.767879 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" podStartSLOduration=2.394287795 podStartE2EDuration="2.767861422s" podCreationTimestamp="2025-10-03 15:59:26 +0000 UTC" firstStartedPulling="2025-10-03 15:59:27.752599228 +0000 UTC m=+8936.955942655" lastFinishedPulling="2025-10-03 15:59:28.126172865 +0000 UTC m=+8937.329516282" observedRunningTime="2025-10-03 15:59:28.761949708 +0000 UTC m=+8937.965293125" watchObservedRunningTime="2025-10-03 15:59:28.767861422 +0000 UTC m=+8937.971204839" Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.045187 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.046011 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.046089 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.047154 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.047233 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78" gracePeriod=600 Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.851744 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78" exitCode=0 Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.851853 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78"} Oct 03 15:59:36 crc kubenswrapper[4959]: I1003 15:59:36.852346 4959 scope.go:117] "RemoveContainer" containerID="c977985bb5945320e79ecad86948e6aac7d95f11bebad6a93a0c40d57a241236" Oct 03 15:59:37 crc kubenswrapper[4959]: I1003 15:59:37.869914 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093"} Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.188720 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj"] Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.192365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.194488 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.194692 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.224857 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj"] Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.370915 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.371089 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.371215 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnwwb\" (UniqueName: \"kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.473825 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.473997 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.474114 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnwwb\" (UniqueName: \"kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.474744 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.482816 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.506578 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnwwb\" (UniqueName: \"kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb\") pod \"collect-profiles-29325120-vsrtj\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:00 crc kubenswrapper[4959]: I1003 16:00:00.530480 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:01 crc kubenswrapper[4959]: I1003 16:00:01.044743 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj"] Oct 03 16:00:01 crc kubenswrapper[4959]: I1003 16:00:01.190030 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" event={"ID":"3423e169-747e-400d-b18e-a947bc3667bc","Type":"ContainerStarted","Data":"0ef291c2f71eabb2d03d16e12a19374c77e48a8baca290a81506e8fda47c43d3"} Oct 03 16:00:02 crc kubenswrapper[4959]: I1003 16:00:02.203768 4959 generic.go:334] "Generic (PLEG): container finished" podID="3423e169-747e-400d-b18e-a947bc3667bc" containerID="3e25531ca42bcc658cef7689c4098387d8d5bd09bed207161c21d30a6c5db0a8" exitCode=0 Oct 03 16:00:02 crc kubenswrapper[4959]: I1003 16:00:02.203838 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" event={"ID":"3423e169-747e-400d-b18e-a947bc3667bc","Type":"ContainerDied","Data":"3e25531ca42bcc658cef7689c4098387d8d5bd09bed207161c21d30a6c5db0a8"} Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.602964 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.745704 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnwwb\" (UniqueName: \"kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb\") pod \"3423e169-747e-400d-b18e-a947bc3667bc\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.745822 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume\") pod \"3423e169-747e-400d-b18e-a947bc3667bc\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.746094 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume\") pod \"3423e169-747e-400d-b18e-a947bc3667bc\" (UID: \"3423e169-747e-400d-b18e-a947bc3667bc\") " Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.746591 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "3423e169-747e-400d-b18e-a947bc3667bc" (UID: "3423e169-747e-400d-b18e-a947bc3667bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.747224 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3423e169-747e-400d-b18e-a947bc3667bc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.763400 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb" (OuterVolumeSpecName: "kube-api-access-vnwwb") pod "3423e169-747e-400d-b18e-a947bc3667bc" (UID: "3423e169-747e-400d-b18e-a947bc3667bc"). InnerVolumeSpecName "kube-api-access-vnwwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.764352 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3423e169-747e-400d-b18e-a947bc3667bc" (UID: "3423e169-747e-400d-b18e-a947bc3667bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.849881 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3423e169-747e-400d-b18e-a947bc3667bc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:00:03 crc kubenswrapper[4959]: I1003 16:00:03.850220 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnwwb\" (UniqueName: \"kubernetes.io/projected/3423e169-747e-400d-b18e-a947bc3667bc-kube-api-access-vnwwb\") on node \"crc\" DevicePath \"\"" Oct 03 16:00:04 crc kubenswrapper[4959]: I1003 16:00:04.227365 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" event={"ID":"3423e169-747e-400d-b18e-a947bc3667bc","Type":"ContainerDied","Data":"0ef291c2f71eabb2d03d16e12a19374c77e48a8baca290a81506e8fda47c43d3"} Oct 03 16:00:04 crc kubenswrapper[4959]: I1003 16:00:04.227404 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ef291c2f71eabb2d03d16e12a19374c77e48a8baca290a81506e8fda47c43d3" Oct 03 16:00:04 crc kubenswrapper[4959]: I1003 16:00:04.227419 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325120-vsrtj" Oct 03 16:00:04 crc kubenswrapper[4959]: I1003 16:00:04.674701 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8"] Oct 03 16:00:04 crc kubenswrapper[4959]: I1003 16:00:04.684102 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325075-29nr8"] Oct 03 16:00:05 crc kubenswrapper[4959]: I1003 16:00:05.706921 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1833901d-4e6a-495f-a1d4-e6b0221acdd8" path="/var/lib/kubelet/pods/1833901d-4e6a-495f-a1d4-e6b0221acdd8/volumes" Oct 03 16:00:12 crc kubenswrapper[4959]: I1003 16:00:12.854292 4959 scope.go:117] "RemoveContainer" containerID="dc269c41f329990be68dd79a167c80089053f01589762f1624ac83de8de37b29" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.191768 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325121-h72rz"] Oct 03 16:01:00 crc kubenswrapper[4959]: E1003 16:01:00.193061 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3423e169-747e-400d-b18e-a947bc3667bc" containerName="collect-profiles" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.193084 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3423e169-747e-400d-b18e-a947bc3667bc" containerName="collect-profiles" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.193532 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3423e169-747e-400d-b18e-a947bc3667bc" containerName="collect-profiles" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.194833 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.210029 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325121-h72rz"] Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.332339 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.332688 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.332785 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvd84\" (UniqueName: \"kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.332834 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.435121 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.435365 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.435556 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.435676 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvd84\" (UniqueName: \"kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.442642 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.444742 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.448099 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.465712 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvd84\" (UniqueName: \"kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84\") pod \"keystone-cron-29325121-h72rz\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:00 crc kubenswrapper[4959]: I1003 16:01:00.527941 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:01 crc kubenswrapper[4959]: I1003 16:01:01.109381 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325121-h72rz"] Oct 03 16:01:02 crc kubenswrapper[4959]: I1003 16:01:02.011368 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325121-h72rz" event={"ID":"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13","Type":"ContainerStarted","Data":"3450e8006856c700bd9ecb73324bfc9be7119eafa02720af85a0b4d39e15bb4f"} Oct 03 16:01:02 crc kubenswrapper[4959]: I1003 16:01:02.012042 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325121-h72rz" event={"ID":"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13","Type":"ContainerStarted","Data":"11260da0151fdee4397132ba9ec65d0a527d48ead985fac21a4b15250f4806c6"} Oct 03 16:01:02 crc kubenswrapper[4959]: I1003 16:01:02.055822 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325121-h72rz" podStartSLOduration=2.055789051 podStartE2EDuration="2.055789051s" podCreationTimestamp="2025-10-03 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:01:02.040621651 +0000 UTC m=+9031.243965138" watchObservedRunningTime="2025-10-03 16:01:02.055789051 +0000 UTC m=+9031.259132508" Oct 03 16:01:04 crc kubenswrapper[4959]: I1003 16:01:04.038634 4959 generic.go:334] "Generic (PLEG): container finished" podID="3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" containerID="3450e8006856c700bd9ecb73324bfc9be7119eafa02720af85a0b4d39e15bb4f" exitCode=0 Oct 03 16:01:04 crc kubenswrapper[4959]: I1003 16:01:04.038672 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325121-h72rz" event={"ID":"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13","Type":"ContainerDied","Data":"3450e8006856c700bd9ecb73324bfc9be7119eafa02720af85a0b4d39e15bb4f"} Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.605701 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.672249 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvd84\" (UniqueName: \"kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84\") pod \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.672366 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data\") pod \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.672450 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys\") pod \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.672660 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle\") pod \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\" (UID: \"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13\") " Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.681718 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" (UID: "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.683403 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84" (OuterVolumeSpecName: "kube-api-access-mvd84") pod "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" (UID: "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13"). InnerVolumeSpecName "kube-api-access-mvd84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.710883 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" (UID: "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.752396 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data" (OuterVolumeSpecName: "config-data") pod "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" (UID: "3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.775933 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvd84\" (UniqueName: \"kubernetes.io/projected/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-kube-api-access-mvd84\") on node \"crc\" DevicePath \"\"" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.776002 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.776020 4959 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 16:01:05 crc kubenswrapper[4959]: I1003 16:01:05.776041 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:01:06 crc kubenswrapper[4959]: I1003 16:01:06.066265 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325121-h72rz" event={"ID":"3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13","Type":"ContainerDied","Data":"11260da0151fdee4397132ba9ec65d0a527d48ead985fac21a4b15250f4806c6"} Oct 03 16:01:06 crc kubenswrapper[4959]: I1003 16:01:06.066317 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11260da0151fdee4397132ba9ec65d0a527d48ead985fac21a4b15250f4806c6" Oct 03 16:01:06 crc kubenswrapper[4959]: I1003 16:01:06.066351 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325121-h72rz" Oct 03 16:02:06 crc kubenswrapper[4959]: I1003 16:02:06.044394 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:02:06 crc kubenswrapper[4959]: I1003 16:02:06.044974 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:02:36 crc kubenswrapper[4959]: I1003 16:02:36.045418 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:02:36 crc kubenswrapper[4959]: I1003 16:02:36.046011 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:03:06 crc kubenswrapper[4959]: I1003 16:03:06.044606 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:03:06 crc kubenswrapper[4959]: I1003 16:03:06.045271 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:03:06 crc kubenswrapper[4959]: I1003 16:03:06.045400 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 16:03:06 crc kubenswrapper[4959]: I1003 16:03:06.586886 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:03:06 crc kubenswrapper[4959]: I1003 16:03:06.586956 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" gracePeriod=600 Oct 03 16:03:07 crc kubenswrapper[4959]: E1003 16:03:07.392634 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:07 crc kubenswrapper[4959]: I1003 16:03:07.600370 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" exitCode=0 Oct 03 16:03:07 crc kubenswrapper[4959]: I1003 16:03:07.600424 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093"} Oct 03 16:03:07 crc kubenswrapper[4959]: I1003 16:03:07.600469 4959 scope.go:117] "RemoveContainer" containerID="a9da077fd3a4dd8e356fbe3f4343b59420379dffabd2a9e42ca335fdde6bcc78" Oct 03 16:03:07 crc kubenswrapper[4959]: I1003 16:03:07.601058 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:03:07 crc kubenswrapper[4959]: E1003 16:03:07.601441 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:18 crc kubenswrapper[4959]: I1003 16:03:18.686995 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:03:18 crc kubenswrapper[4959]: E1003 16:03:18.688282 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:29 crc kubenswrapper[4959]: I1003 16:03:29.685566 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:03:29 crc kubenswrapper[4959]: E1003 16:03:29.686415 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:40 crc kubenswrapper[4959]: I1003 16:03:40.686160 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:03:40 crc kubenswrapper[4959]: E1003 16:03:40.687680 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:44 crc kubenswrapper[4959]: I1003 16:03:44.078985 4959 generic.go:334] "Generic (PLEG): container finished" podID="2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" containerID="287bfe353c16a7a1ea2e5a3d4939b0186094e3512f025c5cda2b9ca6e498bd21" exitCode=0 Oct 03 16:03:44 crc kubenswrapper[4959]: I1003 16:03:44.079113 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" event={"ID":"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4","Type":"ContainerDied","Data":"287bfe353c16a7a1ea2e5a3d4939b0186094e3512f025c5cda2b9ca6e498bd21"} Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.588769 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.643608 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.643689 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.643862 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rvhx\" (UniqueName: \"kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.643945 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.644010 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.644110 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0\") pod \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\" (UID: \"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4\") " Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.649481 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph" (OuterVolumeSpecName: "ceph") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.650404 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.653509 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx" (OuterVolumeSpecName: "kube-api-access-6rvhx") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "kube-api-access-6rvhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.677176 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.678023 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.694129 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory" (OuterVolumeSpecName: "inventory") pod "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" (UID: "2eba6fae-31eb-4f5f-9d23-68f490fe5fa4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746409 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rvhx\" (UniqueName: \"kubernetes.io/projected/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-kube-api-access-6rvhx\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746437 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746452 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746465 4959 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746477 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:45 crc kubenswrapper[4959]: I1003 16:03:45.746489 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2eba6fae-31eb-4f5f-9d23-68f490fe5fa4-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:03:46 crc kubenswrapper[4959]: I1003 16:03:46.104732 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" event={"ID":"2eba6fae-31eb-4f5f-9d23-68f490fe5fa4","Type":"ContainerDied","Data":"c2f6ef76549236cf9d4ad94c8053de2ebb37528670e15267897bc349fb8679a0"} Oct 03 16:03:46 crc kubenswrapper[4959]: I1003 16:03:46.105266 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f6ef76549236cf9d4ad94c8053de2ebb37528670e15267897bc349fb8679a0" Oct 03 16:03:46 crc kubenswrapper[4959]: I1003 16:03:46.104810 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2c2rq" Oct 03 16:03:51 crc kubenswrapper[4959]: I1003 16:03:51.706073 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:03:51 crc kubenswrapper[4959]: E1003 16:03:51.707045 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:03:59 crc kubenswrapper[4959]: E1003 16:03:59.108061 4959 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.245:47884->38.129.56.245:39745: write tcp 38.129.56.245:47884->38.129.56.245:39745: write: broken pipe Oct 03 16:04:00 crc kubenswrapper[4959]: I1003 16:04:00.441948 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:00 crc kubenswrapper[4959]: I1003 16:04:00.442744 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" containerName="nova-cell0-conductor-conductor" containerID="cri-o://7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a" gracePeriod=30 Oct 03 16:04:00 crc kubenswrapper[4959]: I1003 16:04:00.458304 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:00 crc kubenswrapper[4959]: I1003 16:04:00.458812 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" gracePeriod=30 Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.369986 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq"] Oct 03 16:04:01 crc kubenswrapper[4959]: E1003 16:04:01.371442 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" containerName="keystone-cron" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.371489 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" containerName="keystone-cron" Oct 03 16:04:01 crc kubenswrapper[4959]: E1003 16:04:01.371577 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.371587 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.371970 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eba6fae-31eb-4f5f-9d23-68f490fe5fa4" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.372020 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13" containerName="keystone-cron" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.373130 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.377722 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.377848 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.377970 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.377974 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.378084 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.378298 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-w95dr" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.378548 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.391594 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq"] Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516139 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516182 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516496 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516561 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516636 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516705 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516732 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516790 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516827 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516930 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm8f7\" (UniqueName: \"kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.516984 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619285 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619324 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm8f7\" (UniqueName: \"kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619357 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619414 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619432 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619505 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619563 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619596 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.619611 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.625795 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.635521 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.635795 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.636395 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.637367 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.637564 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.646166 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.647279 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.651024 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.651842 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.672604 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm8f7\" (UniqueName: \"kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.710917 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.773128 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.773590 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-log" containerID="cri-o://bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5" gracePeriod=30 Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.774170 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-api" containerID="cri-o://a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e" gracePeriod=30 Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.829679 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.829914 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerName="nova-scheduler-scheduler" containerID="cri-o://2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" gracePeriod=30 Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.877040 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.877302 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-log" containerID="cri-o://11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1" gracePeriod=30 Oct 03 16:04:01 crc kubenswrapper[4959]: I1003 16:04:01.877812 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-metadata" containerID="cri-o://73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b" gracePeriod=30 Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.290115 4959 generic.go:334] "Generic (PLEG): container finished" podID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerID="11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1" exitCode=143 Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.290209 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerDied","Data":"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1"} Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.292683 4959 generic.go:334] "Generic (PLEG): container finished" podID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerID="a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" exitCode=0 Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.292740 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d000b1a0-ae4e-4167-aef4-88cffa9b54cd","Type":"ContainerDied","Data":"a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af"} Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.294371 4959 generic.go:334] "Generic (PLEG): container finished" podID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerID="bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5" exitCode=143 Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.294416 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerDied","Data":"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5"} Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.597526 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq"] Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.631494 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:04:02 crc kubenswrapper[4959]: E1003 16:04:02.664940 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af is running failed: container process not found" containerID="a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 16:04:02 crc kubenswrapper[4959]: E1003 16:04:02.665397 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af is running failed: container process not found" containerID="a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 16:04:02 crc kubenswrapper[4959]: E1003 16:04:02.665777 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af is running failed: container process not found" containerID="a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 03 16:04:02 crc kubenswrapper[4959]: E1003 16:04:02.665819 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerName="nova-cell1-conductor-conductor" Oct 03 16:04:02 crc kubenswrapper[4959]: I1003 16:04:02.836602 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.012233 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data\") pod \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.012341 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle\") pod \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.012484 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7w8b\" (UniqueName: \"kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b\") pod \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\" (UID: \"d000b1a0-ae4e-4167-aef4-88cffa9b54cd\") " Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.016946 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b" (OuterVolumeSpecName: "kube-api-access-v7w8b") pod "d000b1a0-ae4e-4167-aef4-88cffa9b54cd" (UID: "d000b1a0-ae4e-4167-aef4-88cffa9b54cd"). InnerVolumeSpecName "kube-api-access-v7w8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.066986 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d000b1a0-ae4e-4167-aef4-88cffa9b54cd" (UID: "d000b1a0-ae4e-4167-aef4-88cffa9b54cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.083071 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data" (OuterVolumeSpecName: "config-data") pod "d000b1a0-ae4e-4167-aef4-88cffa9b54cd" (UID: "d000b1a0-ae4e-4167-aef4-88cffa9b54cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.115292 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7w8b\" (UniqueName: \"kubernetes.io/projected/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-kube-api-access-v7w8b\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.115339 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.115349 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d000b1a0-ae4e-4167-aef4-88cffa9b54cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.308008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d000b1a0-ae4e-4167-aef4-88cffa9b54cd","Type":"ContainerDied","Data":"00689e23c88b06113f86896cdcad526bf420835a403cebaf25793d6a09167c94"} Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.308073 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.308090 4959 scope.go:117] "RemoveContainer" containerID="a44244aca00c84df28ab816808a0eff79e56f6a172bd90e27995f8c54b76d9af" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.312025 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" event={"ID":"338c9aed-0345-4607-b087-b2523f403bd4","Type":"ContainerStarted","Data":"eeee0271ab54afbb3a094abb28b512877a20c1395697aae5b7f794a9de1acb8e"} Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.312054 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" event={"ID":"338c9aed-0345-4607-b087-b2523f403bd4","Type":"ContainerStarted","Data":"2714a67aafc3668538a8dc8d50d49ce8656343a1a98d8d9e1f7acca6aff508d7"} Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.345753 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" podStartSLOduration=2.156249023 podStartE2EDuration="2.345726744s" podCreationTimestamp="2025-10-03 16:04:01 +0000 UTC" firstStartedPulling="2025-10-03 16:04:02.630461769 +0000 UTC m=+9211.833805186" lastFinishedPulling="2025-10-03 16:04:02.81993949 +0000 UTC m=+9212.023282907" observedRunningTime="2025-10-03 16:04:03.333946658 +0000 UTC m=+9212.537290075" watchObservedRunningTime="2025-10-03 16:04:03.345726744 +0000 UTC m=+9212.549070171" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.364655 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.380286 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.396767 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:03 crc kubenswrapper[4959]: E1003 16:04:03.397319 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerName="nova-cell1-conductor-conductor" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.397336 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerName="nova-cell1-conductor-conductor" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.397544 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" containerName="nova-cell1-conductor-conductor" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.398398 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.400935 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.408521 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.523865 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.523930 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.524126 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p4xm\" (UniqueName: \"kubernetes.io/projected/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-kube-api-access-9p4xm\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.625644 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.625691 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.625789 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p4xm\" (UniqueName: \"kubernetes.io/projected/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-kube-api-access-9p4xm\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.630534 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.630926 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.645870 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p4xm\" (UniqueName: \"kubernetes.io/projected/4e9434fc-ae1c-42e7-95de-3735c19e0c1f-kube-api-access-9p4xm\") pod \"nova-cell1-conductor-0\" (UID: \"4e9434fc-ae1c-42e7-95de-3735c19e0c1f\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.697797 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d000b1a0-ae4e-4167-aef4-88cffa9b54cd" path="/var/lib/kubelet/pods/d000b1a0-ae4e-4167-aef4-88cffa9b54cd/volumes" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.713365 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:03 crc kubenswrapper[4959]: E1003 16:04:03.725853 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddc10e1_3a1b_4abf_8277_742f68ab96c2.slice/crio-conmon-7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:04:03 crc kubenswrapper[4959]: I1003 16:04:03.962688 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.135436 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lflfp\" (UniqueName: \"kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp\") pod \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.135606 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle\") pod \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.135683 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data\") pod \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\" (UID: \"fddc10e1-3a1b-4abf-8277-742f68ab96c2\") " Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.143941 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp" (OuterVolumeSpecName: "kube-api-access-lflfp") pod "fddc10e1-3a1b-4abf-8277-742f68ab96c2" (UID: "fddc10e1-3a1b-4abf-8277-742f68ab96c2"). InnerVolumeSpecName "kube-api-access-lflfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.168111 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fddc10e1-3a1b-4abf-8277-742f68ab96c2" (UID: "fddc10e1-3a1b-4abf-8277-742f68ab96c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.181780 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data" (OuterVolumeSpecName: "config-data") pod "fddc10e1-3a1b-4abf-8277-742f68ab96c2" (UID: "fddc10e1-3a1b-4abf-8277-742f68ab96c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.238287 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lflfp\" (UniqueName: \"kubernetes.io/projected/fddc10e1-3a1b-4abf-8277-742f68ab96c2-kube-api-access-lflfp\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.238313 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.238322 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddc10e1-3a1b-4abf-8277-742f68ab96c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.239992 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:04:04 crc kubenswrapper[4959]: W1003 16:04:04.243920 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e9434fc_ae1c_42e7_95de_3735c19e0c1f.slice/crio-dd21f571b4773d36c0127e9cde637096bd46f88ca22d7177c8a598ae18c51c5c WatchSource:0}: Error finding container dd21f571b4773d36c0127e9cde637096bd46f88ca22d7177c8a598ae18c51c5c: Status 404 returned error can't find the container with id dd21f571b4773d36c0127e9cde637096bd46f88ca22d7177c8a598ae18c51c5c Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.329807 4959 generic.go:334] "Generic (PLEG): container finished" podID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" containerID="7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a" exitCode=0 Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.329855 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fddc10e1-3a1b-4abf-8277-742f68ab96c2","Type":"ContainerDied","Data":"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a"} Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.330273 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fddc10e1-3a1b-4abf-8277-742f68ab96c2","Type":"ContainerDied","Data":"4f5e49c44658be4208aa6851cce254fc23a3d9c54100677fa7d4c715dc102d6d"} Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.330292 4959 scope.go:117] "RemoveContainer" containerID="7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.329938 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.338309 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e9434fc-ae1c-42e7-95de-3735c19e0c1f","Type":"ContainerStarted","Data":"dd21f571b4773d36c0127e9cde637096bd46f88ca22d7177c8a598ae18c51c5c"} Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.370549 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.384942 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.389843 4959 scope.go:117] "RemoveContainer" containerID="7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a" Oct 03 16:04:04 crc kubenswrapper[4959]: E1003 16:04:04.394221 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a\": container with ID starting with 7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a not found: ID does not exist" containerID="7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.394284 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a"} err="failed to get container status \"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a\": rpc error: code = NotFound desc = could not find container \"7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a\": container with ID starting with 7379dbd2f3d405d27f7a54d49a43261426fa0c1a215212a2aa22f8027376321a not found: ID does not exist" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.397056 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:04 crc kubenswrapper[4959]: E1003 16:04:04.397610 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" containerName="nova-cell0-conductor-conductor" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.397628 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" containerName="nova-cell0-conductor-conductor" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.397845 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" containerName="nova-cell0-conductor-conductor" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.398757 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.400871 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.409021 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.544234 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.544311 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.544377 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjznz\" (UniqueName: \"kubernetes.io/projected/f24ab41d-8257-4db6-8c24-44a157bfa8b7-kube-api-access-gjznz\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.646476 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.646566 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.646639 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjznz\" (UniqueName: \"kubernetes.io/projected/f24ab41d-8257-4db6-8c24-44a157bfa8b7-kube-api-access-gjznz\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.654909 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.654950 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f24ab41d-8257-4db6-8c24-44a157bfa8b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.666092 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjznz\" (UniqueName: \"kubernetes.io/projected/f24ab41d-8257-4db6-8c24-44a157bfa8b7-kube-api-access-gjznz\") pod \"nova-cell0-conductor-0\" (UID: \"f24ab41d-8257-4db6-8c24-44a157bfa8b7\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.686054 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:04:04 crc kubenswrapper[4959]: E1003 16:04:04.686393 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:04:04 crc kubenswrapper[4959]: I1003 16:04:04.727557 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:05 crc kubenswrapper[4959]: W1003 16:04:05.201041 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf24ab41d_8257_4db6_8c24_44a157bfa8b7.slice/crio-bfd1741f6ffc7d4d5b570fb216ebd873f10b375746d1d3208fc4f8d0ac1f4007 WatchSource:0}: Error finding container bfd1741f6ffc7d4d5b570fb216ebd873f10b375746d1d3208fc4f8d0ac1f4007: Status 404 returned error can't find the container with id bfd1741f6ffc7d4d5b570fb216ebd873f10b375746d1d3208fc4f8d0ac1f4007 Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.206054 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.352347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f24ab41d-8257-4db6-8c24-44a157bfa8b7","Type":"ContainerStarted","Data":"bfd1741f6ffc7d4d5b570fb216ebd873f10b375746d1d3208fc4f8d0ac1f4007"} Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.356226 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4e9434fc-ae1c-42e7-95de-3735c19e0c1f","Type":"ContainerStarted","Data":"a7521c5edf68e2aae3083d89e980cd61651f703cba728333717d9a3a5770d0d9"} Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.356475 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.382248 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.382230122 podStartE2EDuration="2.382230122s" podCreationTimestamp="2025-10-03 16:04:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:04:05.381354531 +0000 UTC m=+9214.584697958" watchObservedRunningTime="2025-10-03 16:04:05.382230122 +0000 UTC m=+9214.585573539" Oct 03 16:04:05 crc kubenswrapper[4959]: I1003 16:04:05.711689 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddc10e1-3a1b-4abf-8277-742f68ab96c2" path="/var/lib/kubelet/pods/fddc10e1-3a1b-4abf-8277-742f68ab96c2/volumes" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.013670 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.098031 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198336 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs\") pod \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198466 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz6ql\" (UniqueName: \"kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql\") pod \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198517 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs\") pod \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198669 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data\") pod \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198703 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data\") pod \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198742 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdrtx\" (UniqueName: \"kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx\") pod \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198775 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle\") pod \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\" (UID: \"82ef8e5b-0c74-4346-b15a-7f663f0ac040\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198825 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle\") pod \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\" (UID: \"4a38ee4c-2ccc-4c8c-a050-204455f015fe\") " Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.198872 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs" (OuterVolumeSpecName: "logs") pod "82ef8e5b-0c74-4346-b15a-7f663f0ac040" (UID: "82ef8e5b-0c74-4346-b15a-7f663f0ac040"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.199259 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ef8e5b-0c74-4346-b15a-7f663f0ac040-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.209123 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx" (OuterVolumeSpecName: "kube-api-access-cdrtx") pod "82ef8e5b-0c74-4346-b15a-7f663f0ac040" (UID: "82ef8e5b-0c74-4346-b15a-7f663f0ac040"). InnerVolumeSpecName "kube-api-access-cdrtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.212303 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs" (OuterVolumeSpecName: "logs") pod "4a38ee4c-2ccc-4c8c-a050-204455f015fe" (UID: "4a38ee4c-2ccc-4c8c-a050-204455f015fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.217695 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql" (OuterVolumeSpecName: "kube-api-access-wz6ql") pod "4a38ee4c-2ccc-4c8c-a050-204455f015fe" (UID: "4a38ee4c-2ccc-4c8c-a050-204455f015fe"). InnerVolumeSpecName "kube-api-access-wz6ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.279507 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data" (OuterVolumeSpecName: "config-data") pod "4a38ee4c-2ccc-4c8c-a050-204455f015fe" (UID: "4a38ee4c-2ccc-4c8c-a050-204455f015fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.280348 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82ef8e5b-0c74-4346-b15a-7f663f0ac040" (UID: "82ef8e5b-0c74-4346-b15a-7f663f0ac040"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.306851 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz6ql\" (UniqueName: \"kubernetes.io/projected/4a38ee4c-2ccc-4c8c-a050-204455f015fe-kube-api-access-wz6ql\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.306919 4959 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a38ee4c-2ccc-4c8c-a050-204455f015fe-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.306932 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.306944 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdrtx\" (UniqueName: \"kubernetes.io/projected/82ef8e5b-0c74-4346-b15a-7f663f0ac040-kube-api-access-cdrtx\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.306958 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.390709 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f24ab41d-8257-4db6-8c24-44a157bfa8b7","Type":"ContainerStarted","Data":"b03924cabfa85f8f4231d6ba2cbe928e15769af195c6338f49872b39f4ee3651"} Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.392157 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.400744 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data" (OuterVolumeSpecName: "config-data") pod "82ef8e5b-0c74-4346-b15a-7f663f0ac040" (UID: "82ef8e5b-0c74-4346-b15a-7f663f0ac040"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.404424 4959 generic.go:334] "Generic (PLEG): container finished" podID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerID="a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e" exitCode=0 Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.404488 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerDied","Data":"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e"} Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.404514 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4a38ee4c-2ccc-4c8c-a050-204455f015fe","Type":"ContainerDied","Data":"0ed501ea1ef3f8b2623b3ea1992bb9a57c4f79de996bdb9e0dc10b479cd7c23a"} Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.404529 4959 scope.go:117] "RemoveContainer" containerID="a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.404624 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.409793 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ef8e5b-0c74-4346-b15a-7f663f0ac040-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.411360 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a38ee4c-2ccc-4c8c-a050-204455f015fe" (UID: "4a38ee4c-2ccc-4c8c-a050-204455f015fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.435327 4959 generic.go:334] "Generic (PLEG): container finished" podID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerID="73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b" exitCode=0 Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.435574 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.435564 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerDied","Data":"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b"} Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.435767 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ef8e5b-0c74-4346-b15a-7f663f0ac040","Type":"ContainerDied","Data":"7c7c336e549e676d3de05ff33ea5fb5477883424c37e80ae5f1daf2bf415a5a6"} Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.447862 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.447837563 podStartE2EDuration="2.447837563s" podCreationTimestamp="2025-10-03 16:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:04:06.405695428 +0000 UTC m=+9215.609038845" watchObservedRunningTime="2025-10-03 16:04:06.447837563 +0000 UTC m=+9215.651180990" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.466846 4959 scope.go:117] "RemoveContainer" containerID="bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.495287 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.511217 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a38ee4c-2ccc-4c8c-a050-204455f015fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.546013 4959 scope.go:117] "RemoveContainer" containerID="a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.547302 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.547981 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e\": container with ID starting with a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e not found: ID does not exist" containerID="a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.548557 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e"} err="failed to get container status \"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e\": rpc error: code = NotFound desc = could not find container \"a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e\": container with ID starting with a1a87f1f13d88754bac0f11ccd13d5a8e32892b4044f2384df2f877e23b9069e not found: ID does not exist" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.548821 4959 scope.go:117] "RemoveContainer" containerID="bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.549659 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5\": container with ID starting with bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5 not found: ID does not exist" containerID="bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.549778 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5"} err="failed to get container status \"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5\": rpc error: code = NotFound desc = could not find container \"bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5\": container with ID starting with bd5282ca0d9b3b6f75f7f7d5df5c4990cb132074cb67dd32dda2dbc782b922b5 not found: ID does not exist" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.549861 4959 scope.go:117] "RemoveContainer" containerID="73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.587472 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.588289 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-api" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.588315 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-api" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.588358 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-metadata" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.588369 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-metadata" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.588387 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-log" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.588394 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-log" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.588405 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-log" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.588413 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-log" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.590580 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-api" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.590807 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-log" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.590896 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" containerName="nova-api-log" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.591015 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" containerName="nova-metadata-metadata" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.592422 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.599091 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.604214 4959 scope.go:117] "RemoveContainer" containerID="11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.628656 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.645628 4959 scope.go:117] "RemoveContainer" containerID="73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.646008 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b\": container with ID starting with 73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b not found: ID does not exist" containerID="73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.646046 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b"} err="failed to get container status \"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b\": rpc error: code = NotFound desc = could not find container \"73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b\": container with ID starting with 73f0b5f045db6a23d1e168523f851432cd9053bd4afa2694d34f82602810c75b not found: ID does not exist" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.646073 4959 scope.go:117] "RemoveContainer" containerID="11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.646482 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1\": container with ID starting with 11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1 not found: ID does not exist" containerID="11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.646501 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1"} err="failed to get container status \"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1\": rpc error: code = NotFound desc = could not find container \"11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1\": container with ID starting with 11a1f9e26e4fbecbd1c3907ec51f37ef9b5d7017e02155b8b890284c2f61bab1 not found: ID does not exist" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.721811 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.722577 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-config-data\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.722791 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2b08809-1a97-4f9c-835f-0f794d9adf09-logs\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.722873 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8qsh\" (UniqueName: \"kubernetes.io/projected/f2b08809-1a97-4f9c-835f-0f794d9adf09-kube-api-access-r8qsh\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.758955 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 is running failed: container process not found" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.759234 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 is running failed: container process not found" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.759440 4959 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 is running failed: container process not found" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 16:04:06 crc kubenswrapper[4959]: E1003 16:04:06.759479 4959 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerName="nova-scheduler-scheduler" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.801273 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.821655 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.847706 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8qsh\" (UniqueName: \"kubernetes.io/projected/f2b08809-1a97-4f9c-835f-0f794d9adf09-kube-api-access-r8qsh\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.853265 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.853399 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-config-data\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.853775 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2b08809-1a97-4f9c-835f-0f794d9adf09-logs\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.859078 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2b08809-1a97-4f9c-835f-0f794d9adf09-logs\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.867154 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-config-data\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.870845 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2b08809-1a97-4f9c-835f-0f794d9adf09-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.873354 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.874839 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8qsh\" (UniqueName: \"kubernetes.io/projected/f2b08809-1a97-4f9c-835f-0f794d9adf09-kube-api-access-r8qsh\") pod \"nova-metadata-0\" (UID: \"f2b08809-1a97-4f9c-835f-0f794d9adf09\") " pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.892765 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.892857 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.896441 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.917013 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.957452 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.957758 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6cnd\" (UniqueName: \"kubernetes.io/projected/e2e48614-e460-4107-8548-78d3720299b7-kube-api-access-j6cnd\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.957833 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-config-data\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:06 crc kubenswrapper[4959]: I1003 16:04:06.957881 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2e48614-e460-4107-8548-78d3720299b7-logs\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.059853 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6cnd\" (UniqueName: \"kubernetes.io/projected/e2e48614-e460-4107-8548-78d3720299b7-kube-api-access-j6cnd\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.060237 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-config-data\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.060277 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2e48614-e460-4107-8548-78d3720299b7-logs\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.060475 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.062277 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2e48614-e460-4107-8548-78d3720299b7-logs\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.069006 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-config-data\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.071059 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.078844 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2e48614-e460-4107-8548-78d3720299b7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.082416 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6cnd\" (UniqueName: \"kubernetes.io/projected/e2e48614-e460-4107-8548-78d3720299b7-kube-api-access-j6cnd\") pod \"nova-api-0\" (UID: \"e2e48614-e460-4107-8548-78d3720299b7\") " pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.161637 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle\") pod \"91f1ca45-586a-4462-85e2-385dccbcc0c1\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.161742 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9jk6\" (UniqueName: \"kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6\") pod \"91f1ca45-586a-4462-85e2-385dccbcc0c1\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.162008 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data\") pod \"91f1ca45-586a-4462-85e2-385dccbcc0c1\" (UID: \"91f1ca45-586a-4462-85e2-385dccbcc0c1\") " Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.168432 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6" (OuterVolumeSpecName: "kube-api-access-j9jk6") pod "91f1ca45-586a-4462-85e2-385dccbcc0c1" (UID: "91f1ca45-586a-4462-85e2-385dccbcc0c1"). InnerVolumeSpecName "kube-api-access-j9jk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.194722 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91f1ca45-586a-4462-85e2-385dccbcc0c1" (UID: "91f1ca45-586a-4462-85e2-385dccbcc0c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.196042 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data" (OuterVolumeSpecName: "config-data") pod "91f1ca45-586a-4462-85e2-385dccbcc0c1" (UID: "91f1ca45-586a-4462-85e2-385dccbcc0c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.265545 4959 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.265579 4959 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f1ca45-586a-4462-85e2-385dccbcc0c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.265592 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9jk6\" (UniqueName: \"kubernetes.io/projected/91f1ca45-586a-4462-85e2-385dccbcc0c1-kube-api-access-j9jk6\") on node \"crc\" DevicePath \"\"" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.362803 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.434796 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:04:07 crc kubenswrapper[4959]: W1003 16:04:07.445860 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2b08809_1a97_4f9c_835f_0f794d9adf09.slice/crio-37bc8b31c93f7ec86702ea85c07dc166e6ba032d146b34770b3484fa4fc22dc0 WatchSource:0}: Error finding container 37bc8b31c93f7ec86702ea85c07dc166e6ba032d146b34770b3484fa4fc22dc0: Status 404 returned error can't find the container with id 37bc8b31c93f7ec86702ea85c07dc166e6ba032d146b34770b3484fa4fc22dc0 Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.447110 4959 generic.go:334] "Generic (PLEG): container finished" podID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" exitCode=0 Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.447165 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91f1ca45-586a-4462-85e2-385dccbcc0c1","Type":"ContainerDied","Data":"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551"} Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.447209 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91f1ca45-586a-4462-85e2-385dccbcc0c1","Type":"ContainerDied","Data":"1d1205c30eb8212e9aa08083d15196b180a7665fc485ea79a098a28afb9ea04f"} Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.447226 4959 scope.go:117] "RemoveContainer" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.447339 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.504472 4959 scope.go:117] "RemoveContainer" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" Oct 03 16:04:07 crc kubenswrapper[4959]: E1003 16:04:07.504988 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551\": container with ID starting with 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 not found: ID does not exist" containerID="2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.505025 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551"} err="failed to get container status \"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551\": rpc error: code = NotFound desc = could not find container \"2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551\": container with ID starting with 2389101f111216d59ceca01918447c9846e1c0efd82b2a6eae81411a4e198551 not found: ID does not exist" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.520336 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.533210 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.541402 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:07 crc kubenswrapper[4959]: E1003 16:04:07.542066 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerName="nova-scheduler-scheduler" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.542093 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerName="nova-scheduler-scheduler" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.542351 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" containerName="nova-scheduler-scheduler" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.543490 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.547676 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.560847 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.571162 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.571328 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-467d5\" (UniqueName: \"kubernetes.io/projected/a18341b8-2b61-438d-afe0-8598f362b247-kube-api-access-467d5\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.572263 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-config-data\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.678523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.679033 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-467d5\" (UniqueName: \"kubernetes.io/projected/a18341b8-2b61-438d-afe0-8598f362b247-kube-api-access-467d5\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.679235 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-config-data\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.686994 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-config-data\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.703400 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a18341b8-2b61-438d-afe0-8598f362b247-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.706342 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-467d5\" (UniqueName: \"kubernetes.io/projected/a18341b8-2b61-438d-afe0-8598f362b247-kube-api-access-467d5\") pod \"nova-scheduler-0\" (UID: \"a18341b8-2b61-438d-afe0-8598f362b247\") " pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.736248 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a38ee4c-2ccc-4c8c-a050-204455f015fe" path="/var/lib/kubelet/pods/4a38ee4c-2ccc-4c8c-a050-204455f015fe/volumes" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.738290 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ef8e5b-0c74-4346-b15a-7f663f0ac040" path="/var/lib/kubelet/pods/82ef8e5b-0c74-4346-b15a-7f663f0ac040/volumes" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.738894 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f1ca45-586a-4462-85e2-385dccbcc0c1" path="/var/lib/kubelet/pods/91f1ca45-586a-4462-85e2-385dccbcc0c1/volumes" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.871384 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:04:07 crc kubenswrapper[4959]: I1003 16:04:07.893421 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.345558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:04:08 crc kubenswrapper[4959]: W1003 16:04:08.351925 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda18341b8_2b61_438d_afe0_8598f362b247.slice/crio-9411f2e3bc221d662ce16b6753c692f75de210b0cabbead835e270c105b91bc6 WatchSource:0}: Error finding container 9411f2e3bc221d662ce16b6753c692f75de210b0cabbead835e270c105b91bc6: Status 404 returned error can't find the container with id 9411f2e3bc221d662ce16b6753c692f75de210b0cabbead835e270c105b91bc6 Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.470950 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e2e48614-e460-4107-8548-78d3720299b7","Type":"ContainerStarted","Data":"668900317abd67fe3e50ddbc012f600ca5216c02879728057ce982a9ea99602c"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.471012 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e2e48614-e460-4107-8548-78d3720299b7","Type":"ContainerStarted","Data":"65232d791ed586b51080ac97c3239273f54045067e701cc9ff710c09eb6f7946"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.471024 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e2e48614-e460-4107-8548-78d3720299b7","Type":"ContainerStarted","Data":"60c72ac9fe513f0a298684ecde1d17d20e38cabbe9344d7f4f8f2d61b0359199"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.482336 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2b08809-1a97-4f9c-835f-0f794d9adf09","Type":"ContainerStarted","Data":"269ab94ead783d9e28959df09771a134e6d9b060e70b436ae99851579ace7ca9"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.482390 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2b08809-1a97-4f9c-835f-0f794d9adf09","Type":"ContainerStarted","Data":"de76197e624abad83697e7567050986f5adaa734310ffdf35a8622afd27733c5"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.482401 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2b08809-1a97-4f9c-835f-0f794d9adf09","Type":"ContainerStarted","Data":"37bc8b31c93f7ec86702ea85c07dc166e6ba032d146b34770b3484fa4fc22dc0"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.490104 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a18341b8-2b61-438d-afe0-8598f362b247","Type":"ContainerStarted","Data":"9411f2e3bc221d662ce16b6753c692f75de210b0cabbead835e270c105b91bc6"} Oct 03 16:04:08 crc kubenswrapper[4959]: I1003 16:04:08.494113 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.494095129 podStartE2EDuration="2.494095129s" podCreationTimestamp="2025-10-03 16:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:04:08.4896401 +0000 UTC m=+9217.692983537" watchObservedRunningTime="2025-10-03 16:04:08.494095129 +0000 UTC m=+9217.697438546" Oct 03 16:04:09 crc kubenswrapper[4959]: I1003 16:04:09.505041 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a18341b8-2b61-438d-afe0-8598f362b247","Type":"ContainerStarted","Data":"dc85d6b64ead0c1ca5fb7b482fe608389e7a2ebbcfa6795a6d6267f94a3f394d"} Oct 03 16:04:09 crc kubenswrapper[4959]: I1003 16:04:09.532262 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.532226151 podStartE2EDuration="3.532226151s" podCreationTimestamp="2025-10-03 16:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:04:08.512734222 +0000 UTC m=+9217.716077639" watchObservedRunningTime="2025-10-03 16:04:09.532226151 +0000 UTC m=+9218.735569608" Oct 03 16:04:09 crc kubenswrapper[4959]: I1003 16:04:09.540944 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.540918172 podStartE2EDuration="2.540918172s" podCreationTimestamp="2025-10-03 16:04:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:04:09.524004761 +0000 UTC m=+9218.727348208" watchObservedRunningTime="2025-10-03 16:04:09.540918172 +0000 UTC m=+9218.744261639" Oct 03 16:04:11 crc kubenswrapper[4959]: I1003 16:04:11.918069 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:04:11 crc kubenswrapper[4959]: I1003 16:04:11.918634 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:04:12 crc kubenswrapper[4959]: I1003 16:04:12.872165 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 16:04:13 crc kubenswrapper[4959]: I1003 16:04:13.751441 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 16:04:14 crc kubenswrapper[4959]: I1003 16:04:14.772299 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 16:04:16 crc kubenswrapper[4959]: I1003 16:04:16.918271 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:04:16 crc kubenswrapper[4959]: I1003 16:04:16.918646 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:04:17 crc kubenswrapper[4959]: I1003 16:04:17.363725 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:04:17 crc kubenswrapper[4959]: I1003 16:04:17.364312 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:04:17 crc kubenswrapper[4959]: I1003 16:04:17.686343 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:04:17 crc kubenswrapper[4959]: E1003 16:04:17.686620 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:04:17 crc kubenswrapper[4959]: I1003 16:04:17.871858 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 16:04:17 crc kubenswrapper[4959]: I1003 16:04:17.930065 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 16:04:18 crc kubenswrapper[4959]: I1003 16:04:18.001492 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f2b08809-1a97-4f9c-835f-0f794d9adf09" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:04:18 crc kubenswrapper[4959]: I1003 16:04:18.001492 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f2b08809-1a97-4f9c-835f-0f794d9adf09" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:04:18 crc kubenswrapper[4959]: I1003 16:04:18.455480 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e2e48614-e460-4107-8548-78d3720299b7" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:04:18 crc kubenswrapper[4959]: I1003 16:04:18.455719 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e2e48614-e460-4107-8548-78d3720299b7" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:04:18 crc kubenswrapper[4959]: I1003 16:04:18.639322 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 16:04:26 crc kubenswrapper[4959]: I1003 16:04:26.920863 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 16:04:26 crc kubenswrapper[4959]: I1003 16:04:26.921566 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 16:04:26 crc kubenswrapper[4959]: I1003 16:04:26.923594 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 16:04:26 crc kubenswrapper[4959]: I1003 16:04:26.925462 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.366996 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.367807 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.368879 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.369142 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.717916 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:04:27 crc kubenswrapper[4959]: I1003 16:04:27.726004 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:04:31 crc kubenswrapper[4959]: I1003 16:04:31.701429 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:04:31 crc kubenswrapper[4959]: E1003 16:04:31.702760 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:04:46 crc kubenswrapper[4959]: I1003 16:04:46.686485 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:04:46 crc kubenswrapper[4959]: E1003 16:04:46.688157 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:04:57 crc kubenswrapper[4959]: I1003 16:04:57.685973 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:04:57 crc kubenswrapper[4959]: E1003 16:04:57.686875 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:05:08 crc kubenswrapper[4959]: I1003 16:05:08.685557 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:05:08 crc kubenswrapper[4959]: E1003 16:05:08.686609 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:05:21 crc kubenswrapper[4959]: I1003 16:05:21.693381 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:05:21 crc kubenswrapper[4959]: E1003 16:05:21.694129 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.263048 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.271482 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.276718 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.328608 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp872\" (UniqueName: \"kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.329023 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.329470 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.431402 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.431470 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.431508 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp872\" (UniqueName: \"kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.431854 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.432080 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.456097 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp872\" (UniqueName: \"kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872\") pod \"redhat-operators-jnzt4\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:28 crc kubenswrapper[4959]: I1003 16:05:28.613795 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:29 crc kubenswrapper[4959]: I1003 16:05:29.119558 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:29 crc kubenswrapper[4959]: I1003 16:05:29.484439 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerStarted","Data":"4b60b4d4f9478f52e9dc94e0fab685a350e7ddbe5d25fc0028a6bce147287331"} Oct 03 16:05:30 crc kubenswrapper[4959]: I1003 16:05:30.497288 4959 generic.go:334] "Generic (PLEG): container finished" podID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerID="03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9" exitCode=0 Oct 03 16:05:30 crc kubenswrapper[4959]: I1003 16:05:30.497391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerDied","Data":"03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9"} Oct 03 16:05:32 crc kubenswrapper[4959]: I1003 16:05:32.524458 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerStarted","Data":"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a"} Oct 03 16:05:36 crc kubenswrapper[4959]: I1003 16:05:36.577145 4959 generic.go:334] "Generic (PLEG): container finished" podID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerID="9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a" exitCode=0 Oct 03 16:05:36 crc kubenswrapper[4959]: I1003 16:05:36.577215 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerDied","Data":"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a"} Oct 03 16:05:36 crc kubenswrapper[4959]: I1003 16:05:36.685258 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:05:36 crc kubenswrapper[4959]: E1003 16:05:36.685515 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:05:37 crc kubenswrapper[4959]: I1003 16:05:37.593689 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerStarted","Data":"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e"} Oct 03 16:05:37 crc kubenswrapper[4959]: I1003 16:05:37.631088 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jnzt4" podStartSLOduration=3.122557098 podStartE2EDuration="9.63105567s" podCreationTimestamp="2025-10-03 16:05:28 +0000 UTC" firstStartedPulling="2025-10-03 16:05:30.500306946 +0000 UTC m=+9299.703650373" lastFinishedPulling="2025-10-03 16:05:37.008805478 +0000 UTC m=+9306.212148945" observedRunningTime="2025-10-03 16:05:37.616027045 +0000 UTC m=+9306.819370472" watchObservedRunningTime="2025-10-03 16:05:37.63105567 +0000 UTC m=+9306.834399107" Oct 03 16:05:38 crc kubenswrapper[4959]: I1003 16:05:38.614234 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:38 crc kubenswrapper[4959]: I1003 16:05:38.614317 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:39 crc kubenswrapper[4959]: I1003 16:05:39.668829 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jnzt4" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="registry-server" probeResult="failure" output=< Oct 03 16:05:39 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:05:39 crc kubenswrapper[4959]: > Oct 03 16:05:48 crc kubenswrapper[4959]: I1003 16:05:48.681422 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:48 crc kubenswrapper[4959]: I1003 16:05:48.771863 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:48 crc kubenswrapper[4959]: I1003 16:05:48.934043 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:49 crc kubenswrapper[4959]: I1003 16:05:49.686264 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:05:49 crc kubenswrapper[4959]: E1003 16:05:49.686903 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:05:49 crc kubenswrapper[4959]: I1003 16:05:49.730127 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jnzt4" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="registry-server" containerID="cri-o://ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e" gracePeriod=2 Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.277679 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.331786 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities\") pod \"5416b544-e6f0-4329-a055-c61a7120b2ed\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.331853 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp872\" (UniqueName: \"kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872\") pod \"5416b544-e6f0-4329-a055-c61a7120b2ed\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.331899 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content\") pod \"5416b544-e6f0-4329-a055-c61a7120b2ed\" (UID: \"5416b544-e6f0-4329-a055-c61a7120b2ed\") " Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.337012 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities" (OuterVolumeSpecName: "utilities") pod "5416b544-e6f0-4329-a055-c61a7120b2ed" (UID: "5416b544-e6f0-4329-a055-c61a7120b2ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.353531 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872" (OuterVolumeSpecName: "kube-api-access-kp872") pod "5416b544-e6f0-4329-a055-c61a7120b2ed" (UID: "5416b544-e6f0-4329-a055-c61a7120b2ed"). InnerVolumeSpecName "kube-api-access-kp872". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.421590 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5416b544-e6f0-4329-a055-c61a7120b2ed" (UID: "5416b544-e6f0-4329-a055-c61a7120b2ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.434550 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.434587 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp872\" (UniqueName: \"kubernetes.io/projected/5416b544-e6f0-4329-a055-c61a7120b2ed-kube-api-access-kp872\") on node \"crc\" DevicePath \"\"" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.434598 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5416b544-e6f0-4329-a055-c61a7120b2ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.743767 4959 generic.go:334] "Generic (PLEG): container finished" podID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerID="ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e" exitCode=0 Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.743827 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnzt4" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.743832 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerDied","Data":"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e"} Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.744004 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnzt4" event={"ID":"5416b544-e6f0-4329-a055-c61a7120b2ed","Type":"ContainerDied","Data":"4b60b4d4f9478f52e9dc94e0fab685a350e7ddbe5d25fc0028a6bce147287331"} Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.744035 4959 scope.go:117] "RemoveContainer" containerID="ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.782471 4959 scope.go:117] "RemoveContainer" containerID="9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a" Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.808846 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:50 crc kubenswrapper[4959]: I1003 16:05:50.821450 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jnzt4"] Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.496518 4959 scope.go:117] "RemoveContainer" containerID="03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.574181 4959 scope.go:117] "RemoveContainer" containerID="ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e" Oct 03 16:05:51 crc kubenswrapper[4959]: E1003 16:05:51.574719 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e\": container with ID starting with ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e not found: ID does not exist" containerID="ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.574773 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e"} err="failed to get container status \"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e\": rpc error: code = NotFound desc = could not find container \"ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e\": container with ID starting with ad8e14ff765bacdcda51491601da52c78476bd2077e6221787d5c29e5799bd8e not found: ID does not exist" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.574807 4959 scope.go:117] "RemoveContainer" containerID="9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a" Oct 03 16:05:51 crc kubenswrapper[4959]: E1003 16:05:51.575279 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a\": container with ID starting with 9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a not found: ID does not exist" containerID="9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.575329 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a"} err="failed to get container status \"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a\": rpc error: code = NotFound desc = could not find container \"9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a\": container with ID starting with 9b7a7bda2aba1ddfe88cd9ff23e788a0b544ad8ffa5da3e9d8fa6256608ffb7a not found: ID does not exist" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.575362 4959 scope.go:117] "RemoveContainer" containerID="03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9" Oct 03 16:05:51 crc kubenswrapper[4959]: E1003 16:05:51.575662 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9\": container with ID starting with 03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9 not found: ID does not exist" containerID="03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.575698 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9"} err="failed to get container status \"03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9\": rpc error: code = NotFound desc = could not find container \"03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9\": container with ID starting with 03643ae02368ba1466d916cf7ba3e5155de5858f6caf264979d3fd958cedfcb9 not found: ID does not exist" Oct 03 16:05:51 crc kubenswrapper[4959]: I1003 16:05:51.701181 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" path="/var/lib/kubelet/pods/5416b544-e6f0-4329-a055-c61a7120b2ed/volumes" Oct 03 16:06:03 crc kubenswrapper[4959]: I1003 16:06:03.685524 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:06:03 crc kubenswrapper[4959]: E1003 16:06:03.686227 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:06:16 crc kubenswrapper[4959]: I1003 16:06:16.686135 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:06:16 crc kubenswrapper[4959]: E1003 16:06:16.687217 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:06:27 crc kubenswrapper[4959]: I1003 16:06:27.686677 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:06:27 crc kubenswrapper[4959]: E1003 16:06:27.688065 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:06:39 crc kubenswrapper[4959]: I1003 16:06:39.685473 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:06:39 crc kubenswrapper[4959]: E1003 16:06:39.686480 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:06:51 crc kubenswrapper[4959]: I1003 16:06:51.702779 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:06:51 crc kubenswrapper[4959]: E1003 16:06:51.703947 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:07:05 crc kubenswrapper[4959]: I1003 16:07:05.685441 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:07:05 crc kubenswrapper[4959]: E1003 16:07:05.686403 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:07:16 crc kubenswrapper[4959]: I1003 16:07:16.685352 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:07:16 crc kubenswrapper[4959]: E1003 16:07:16.686113 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:07:29 crc kubenswrapper[4959]: I1003 16:07:29.685367 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:07:29 crc kubenswrapper[4959]: E1003 16:07:29.688365 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:07:40 crc kubenswrapper[4959]: I1003 16:07:40.685710 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:07:40 crc kubenswrapper[4959]: E1003 16:07:40.686509 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:07:53 crc kubenswrapper[4959]: I1003 16:07:53.686138 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:07:53 crc kubenswrapper[4959]: E1003 16:07:53.687254 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:08:05 crc kubenswrapper[4959]: I1003 16:08:05.686581 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:08:05 crc kubenswrapper[4959]: E1003 16:08:05.687702 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:08:18 crc kubenswrapper[4959]: I1003 16:08:18.685471 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:08:19 crc kubenswrapper[4959]: I1003 16:08:19.458324 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e"} Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.103936 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:09:39 crc kubenswrapper[4959]: E1003 16:09:39.104961 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="registry-server" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.104980 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="registry-server" Oct 03 16:09:39 crc kubenswrapper[4959]: E1003 16:09:39.105012 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="extract-utilities" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.105023 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="extract-utilities" Oct 03 16:09:39 crc kubenswrapper[4959]: E1003 16:09:39.105058 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="extract-content" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.105073 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="extract-content" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.105414 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="5416b544-e6f0-4329-a055-c61a7120b2ed" containerName="registry-server" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.107818 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.121214 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.245148 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9qgv\" (UniqueName: \"kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.245623 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.245832 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.347685 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9qgv\" (UniqueName: \"kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.347819 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.347848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.348388 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.348398 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.370448 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9qgv\" (UniqueName: \"kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv\") pod \"redhat-marketplace-sr6zx\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.455112 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:39 crc kubenswrapper[4959]: I1003 16:09:39.988970 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:09:40 crc kubenswrapper[4959]: I1003 16:09:40.443945 4959 generic.go:334] "Generic (PLEG): container finished" podID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerID="1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b" exitCode=0 Oct 03 16:09:40 crc kubenswrapper[4959]: I1003 16:09:40.444301 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerDied","Data":"1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b"} Oct 03 16:09:40 crc kubenswrapper[4959]: I1003 16:09:40.444364 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerStarted","Data":"79434d29ebc577d2d1a3114971904f3dc6a4955898cc41640f125e25397bc335"} Oct 03 16:09:40 crc kubenswrapper[4959]: I1003 16:09:40.448247 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:09:42 crc kubenswrapper[4959]: I1003 16:09:42.467236 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerStarted","Data":"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a"} Oct 03 16:09:43 crc kubenswrapper[4959]: E1003 16:09:43.864470 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaff7c8e6_6704_44f1_8862_ffd3b86bc12e.slice/crio-f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:09:44 crc kubenswrapper[4959]: I1003 16:09:44.493269 4959 generic.go:334] "Generic (PLEG): container finished" podID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerID="f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a" exitCode=0 Oct 03 16:09:44 crc kubenswrapper[4959]: I1003 16:09:44.493327 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerDied","Data":"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a"} Oct 03 16:09:45 crc kubenswrapper[4959]: I1003 16:09:45.507340 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerStarted","Data":"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360"} Oct 03 16:09:45 crc kubenswrapper[4959]: I1003 16:09:45.533305 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sr6zx" podStartSLOduration=2.092098281 podStartE2EDuration="6.533284206s" podCreationTimestamp="2025-10-03 16:09:39 +0000 UTC" firstStartedPulling="2025-10-03 16:09:40.447970997 +0000 UTC m=+9549.651314414" lastFinishedPulling="2025-10-03 16:09:44.889156912 +0000 UTC m=+9554.092500339" observedRunningTime="2025-10-03 16:09:45.525330492 +0000 UTC m=+9554.728673919" watchObservedRunningTime="2025-10-03 16:09:45.533284206 +0000 UTC m=+9554.736627623" Oct 03 16:09:49 crc kubenswrapper[4959]: I1003 16:09:49.455786 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:49 crc kubenswrapper[4959]: I1003 16:09:49.456408 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:49 crc kubenswrapper[4959]: I1003 16:09:49.510388 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:59 crc kubenswrapper[4959]: I1003 16:09:59.518284 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:09:59 crc kubenswrapper[4959]: I1003 16:09:59.579613 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:09:59 crc kubenswrapper[4959]: I1003 16:09:59.697109 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sr6zx" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="registry-server" containerID="cri-o://85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360" gracePeriod=2 Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.307655 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.439842 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content\") pod \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.439979 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities\") pod \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.440080 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9qgv\" (UniqueName: \"kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv\") pod \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\" (UID: \"aff7c8e6-6704-44f1-8862-ffd3b86bc12e\") " Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.440835 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities" (OuterVolumeSpecName: "utilities") pod "aff7c8e6-6704-44f1-8862-ffd3b86bc12e" (UID: "aff7c8e6-6704-44f1-8862-ffd3b86bc12e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.447624 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv" (OuterVolumeSpecName: "kube-api-access-f9qgv") pod "aff7c8e6-6704-44f1-8862-ffd3b86bc12e" (UID: "aff7c8e6-6704-44f1-8862-ffd3b86bc12e"). InnerVolumeSpecName "kube-api-access-f9qgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.460825 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aff7c8e6-6704-44f1-8862-ffd3b86bc12e" (UID: "aff7c8e6-6704-44f1-8862-ffd3b86bc12e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.546305 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.546355 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9qgv\" (UniqueName: \"kubernetes.io/projected/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-kube-api-access-f9qgv\") on node \"crc\" DevicePath \"\"" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.546373 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff7c8e6-6704-44f1-8862-ffd3b86bc12e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.709265 4959 generic.go:334] "Generic (PLEG): container finished" podID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerID="85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360" exitCode=0 Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.709305 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerDied","Data":"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360"} Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.709365 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sr6zx" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.709414 4959 scope.go:117] "RemoveContainer" containerID="85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.709391 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sr6zx" event={"ID":"aff7c8e6-6704-44f1-8862-ffd3b86bc12e","Type":"ContainerDied","Data":"79434d29ebc577d2d1a3114971904f3dc6a4955898cc41640f125e25397bc335"} Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.759507 4959 scope.go:117] "RemoveContainer" containerID="f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.764212 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.777259 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sr6zx"] Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.801676 4959 scope.go:117] "RemoveContainer" containerID="1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.839020 4959 scope.go:117] "RemoveContainer" containerID="85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360" Oct 03 16:10:00 crc kubenswrapper[4959]: E1003 16:10:00.839524 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360\": container with ID starting with 85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360 not found: ID does not exist" containerID="85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.839561 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360"} err="failed to get container status \"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360\": rpc error: code = NotFound desc = could not find container \"85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360\": container with ID starting with 85df101a98e78ffba1a5265a4c47962d89cccc57c5087fb5d4837b72d5cc1360 not found: ID does not exist" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.839585 4959 scope.go:117] "RemoveContainer" containerID="f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a" Oct 03 16:10:00 crc kubenswrapper[4959]: E1003 16:10:00.840020 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a\": container with ID starting with f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a not found: ID does not exist" containerID="f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.840078 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a"} err="failed to get container status \"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a\": rpc error: code = NotFound desc = could not find container \"f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a\": container with ID starting with f3d4c361e5c102124b683577387ae22fb7c9357f55d47d45a9732786745f3b6a not found: ID does not exist" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.840120 4959 scope.go:117] "RemoveContainer" containerID="1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b" Oct 03 16:10:00 crc kubenswrapper[4959]: E1003 16:10:00.840490 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b\": container with ID starting with 1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b not found: ID does not exist" containerID="1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b" Oct 03 16:10:00 crc kubenswrapper[4959]: I1003 16:10:00.840553 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b"} err="failed to get container status \"1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b\": rpc error: code = NotFound desc = could not find container \"1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b\": container with ID starting with 1f9718b84851d67ceb818358dea37299cb9b09f9d89fbd4e21b774e975e6ca8b not found: ID does not exist" Oct 03 16:10:01 crc kubenswrapper[4959]: I1003 16:10:01.697527 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" path="/var/lib/kubelet/pods/aff7c8e6-6704-44f1-8862-ffd3b86bc12e/volumes" Oct 03 16:10:36 crc kubenswrapper[4959]: I1003 16:10:36.044546 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:10:36 crc kubenswrapper[4959]: I1003 16:10:36.045186 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:02.999706 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:03 crc kubenswrapper[4959]: E1003 16:11:03.000809 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="registry-server" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.000824 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="registry-server" Oct 03 16:11:03 crc kubenswrapper[4959]: E1003 16:11:03.000849 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="extract-utilities" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.000855 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="extract-utilities" Oct 03 16:11:03 crc kubenswrapper[4959]: E1003 16:11:03.000864 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="extract-content" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.000870 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="extract-content" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.001080 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="aff7c8e6-6704-44f1-8862-ffd3b86bc12e" containerName="registry-server" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.002710 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.010655 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.046933 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.047042 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.047096 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c45l\" (UniqueName: \"kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.149294 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c45l\" (UniqueName: \"kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.149545 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.149701 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.150090 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.150169 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.191368 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c45l\" (UniqueName: \"kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l\") pod \"community-operators-7vpwk\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.323931 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:03 crc kubenswrapper[4959]: I1003 16:11:03.830553 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:04 crc kubenswrapper[4959]: I1003 16:11:04.494421 4959 generic.go:334] "Generic (PLEG): container finished" podID="59e33877-c720-4324-ae79-515b18cb0693" containerID="c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a" exitCode=0 Oct 03 16:11:04 crc kubenswrapper[4959]: I1003 16:11:04.494831 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerDied","Data":"c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a"} Oct 03 16:11:04 crc kubenswrapper[4959]: I1003 16:11:04.494865 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerStarted","Data":"3040169d970b3eeb5295918db854dede86f463fdd8186221d742ca93157445e9"} Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.594650 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.597892 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.602452 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.602522 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.602731 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwd79\" (UniqueName: \"kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.609251 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.705106 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.705175 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.705241 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwd79\" (UniqueName: \"kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.705716 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.706051 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.731851 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwd79\" (UniqueName: \"kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79\") pod \"certified-operators-zg5wt\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:05 crc kubenswrapper[4959]: I1003 16:11:05.922007 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:06 crc kubenswrapper[4959]: I1003 16:11:06.047343 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:11:06 crc kubenswrapper[4959]: I1003 16:11:06.047646 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:11:06 crc kubenswrapper[4959]: I1003 16:11:06.517008 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerStarted","Data":"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14"} Oct 03 16:11:06 crc kubenswrapper[4959]: I1003 16:11:06.580113 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:07 crc kubenswrapper[4959]: I1003 16:11:07.534384 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc019303-87b0-499a-86c1-af977b6cce5c" containerID="7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29" exitCode=0 Oct 03 16:11:07 crc kubenswrapper[4959]: I1003 16:11:07.534609 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerDied","Data":"7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29"} Oct 03 16:11:07 crc kubenswrapper[4959]: I1003 16:11:07.535148 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerStarted","Data":"4d26df41022d2f9f673f139ac10f41ee69e7630f318db1fb890da62d43074d12"} Oct 03 16:11:07 crc kubenswrapper[4959]: I1003 16:11:07.538710 4959 generic.go:334] "Generic (PLEG): container finished" podID="59e33877-c720-4324-ae79-515b18cb0693" containerID="2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14" exitCode=0 Oct 03 16:11:07 crc kubenswrapper[4959]: I1003 16:11:07.538763 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerDied","Data":"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14"} Oct 03 16:11:08 crc kubenswrapper[4959]: I1003 16:11:08.550806 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerStarted","Data":"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6"} Oct 03 16:11:08 crc kubenswrapper[4959]: I1003 16:11:08.553440 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerStarted","Data":"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994"} Oct 03 16:11:08 crc kubenswrapper[4959]: I1003 16:11:08.597452 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7vpwk" podStartSLOduration=3.136072145 podStartE2EDuration="6.597433886s" podCreationTimestamp="2025-10-03 16:11:02 +0000 UTC" firstStartedPulling="2025-10-03 16:11:04.501148154 +0000 UTC m=+9633.704491611" lastFinishedPulling="2025-10-03 16:11:07.962509935 +0000 UTC m=+9637.165853352" observedRunningTime="2025-10-03 16:11:08.588831726 +0000 UTC m=+9637.792175153" watchObservedRunningTime="2025-10-03 16:11:08.597433886 +0000 UTC m=+9637.800777303" Oct 03 16:11:10 crc kubenswrapper[4959]: I1003 16:11:10.581449 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc019303-87b0-499a-86c1-af977b6cce5c" containerID="df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6" exitCode=0 Oct 03 16:11:10 crc kubenswrapper[4959]: I1003 16:11:10.581581 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerDied","Data":"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6"} Oct 03 16:11:12 crc kubenswrapper[4959]: I1003 16:11:12.610804 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerStarted","Data":"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32"} Oct 03 16:11:12 crc kubenswrapper[4959]: I1003 16:11:12.648735 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zg5wt" podStartSLOduration=4.009781011 podStartE2EDuration="7.648711613s" podCreationTimestamp="2025-10-03 16:11:05 +0000 UTC" firstStartedPulling="2025-10-03 16:11:07.539296286 +0000 UTC m=+9636.742639723" lastFinishedPulling="2025-10-03 16:11:11.178226898 +0000 UTC m=+9640.381570325" observedRunningTime="2025-10-03 16:11:12.62973834 +0000 UTC m=+9641.833081807" watchObservedRunningTime="2025-10-03 16:11:12.648711613 +0000 UTC m=+9641.852055040" Oct 03 16:11:13 crc kubenswrapper[4959]: I1003 16:11:13.324052 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:13 crc kubenswrapper[4959]: I1003 16:11:13.324106 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:13 crc kubenswrapper[4959]: I1003 16:11:13.635241 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:13 crc kubenswrapper[4959]: I1003 16:11:13.710705 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:14 crc kubenswrapper[4959]: I1003 16:11:14.987209 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:15 crc kubenswrapper[4959]: I1003 16:11:15.660589 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7vpwk" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="registry-server" containerID="cri-o://b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994" gracePeriod=2 Oct 03 16:11:15 crc kubenswrapper[4959]: I1003 16:11:15.922678 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:15 crc kubenswrapper[4959]: I1003 16:11:15.922763 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.222727 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.375941 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities\") pod \"59e33877-c720-4324-ae79-515b18cb0693\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.378035 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content\") pod \"59e33877-c720-4324-ae79-515b18cb0693\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.377039 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities" (OuterVolumeSpecName: "utilities") pod "59e33877-c720-4324-ae79-515b18cb0693" (UID: "59e33877-c720-4324-ae79-515b18cb0693"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.378435 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c45l\" (UniqueName: \"kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l\") pod \"59e33877-c720-4324-ae79-515b18cb0693\" (UID: \"59e33877-c720-4324-ae79-515b18cb0693\") " Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.380060 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.389924 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l" (OuterVolumeSpecName: "kube-api-access-5c45l") pod "59e33877-c720-4324-ae79-515b18cb0693" (UID: "59e33877-c720-4324-ae79-515b18cb0693"). InnerVolumeSpecName "kube-api-access-5c45l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.428957 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59e33877-c720-4324-ae79-515b18cb0693" (UID: "59e33877-c720-4324-ae79-515b18cb0693"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.482869 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59e33877-c720-4324-ae79-515b18cb0693-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.482930 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c45l\" (UniqueName: \"kubernetes.io/projected/59e33877-c720-4324-ae79-515b18cb0693-kube-api-access-5c45l\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.673751 4959 generic.go:334] "Generic (PLEG): container finished" podID="59e33877-c720-4324-ae79-515b18cb0693" containerID="b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994" exitCode=0 Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.673792 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerDied","Data":"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994"} Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.674009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vpwk" event={"ID":"59e33877-c720-4324-ae79-515b18cb0693","Type":"ContainerDied","Data":"3040169d970b3eeb5295918db854dede86f463fdd8186221d742ca93157445e9"} Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.674028 4959 scope.go:117] "RemoveContainer" containerID="b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.674040 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vpwk" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.720489 4959 scope.go:117] "RemoveContainer" containerID="2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.732753 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.745318 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7vpwk"] Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.757695 4959 scope.go:117] "RemoveContainer" containerID="c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.822987 4959 scope.go:117] "RemoveContainer" containerID="b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994" Oct 03 16:11:16 crc kubenswrapper[4959]: E1003 16:11:16.823250 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994\": container with ID starting with b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994 not found: ID does not exist" containerID="b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.823281 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994"} err="failed to get container status \"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994\": rpc error: code = NotFound desc = could not find container \"b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994\": container with ID starting with b2ee1085995b20f217f59adb27f839a31f388357cca0f46670af19480533d994 not found: ID does not exist" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.823305 4959 scope.go:117] "RemoveContainer" containerID="2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14" Oct 03 16:11:16 crc kubenswrapper[4959]: E1003 16:11:16.823676 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14\": container with ID starting with 2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14 not found: ID does not exist" containerID="2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.823711 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14"} err="failed to get container status \"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14\": rpc error: code = NotFound desc = could not find container \"2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14\": container with ID starting with 2c679616d9d376a27a44e9b1e452a1315c8d5359c856b0c76607dbc899d67d14 not found: ID does not exist" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.823733 4959 scope.go:117] "RemoveContainer" containerID="c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a" Oct 03 16:11:16 crc kubenswrapper[4959]: E1003 16:11:16.825046 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a\": container with ID starting with c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a not found: ID does not exist" containerID="c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.825107 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a"} err="failed to get container status \"c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a\": rpc error: code = NotFound desc = could not find container \"c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a\": container with ID starting with c6e217d7b0654390dd84edfab6609d3ac25fff823400343416e4432de577bb8a not found: ID does not exist" Oct 03 16:11:16 crc kubenswrapper[4959]: I1003 16:11:16.983276 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zg5wt" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="registry-server" probeResult="failure" output=< Oct 03 16:11:16 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:11:16 crc kubenswrapper[4959]: > Oct 03 16:11:17 crc kubenswrapper[4959]: I1003 16:11:17.710363 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e33877-c720-4324-ae79-515b18cb0693" path="/var/lib/kubelet/pods/59e33877-c720-4324-ae79-515b18cb0693/volumes" Oct 03 16:11:26 crc kubenswrapper[4959]: I1003 16:11:26.019660 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:26 crc kubenswrapper[4959]: I1003 16:11:26.088476 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:26 crc kubenswrapper[4959]: I1003 16:11:26.267480 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:27 crc kubenswrapper[4959]: I1003 16:11:27.857509 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zg5wt" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="registry-server" containerID="cri-o://6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32" gracePeriod=2 Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.479026 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.582480 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content\") pod \"cc019303-87b0-499a-86c1-af977b6cce5c\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.582542 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwd79\" (UniqueName: \"kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79\") pod \"cc019303-87b0-499a-86c1-af977b6cce5c\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.582616 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities\") pod \"cc019303-87b0-499a-86c1-af977b6cce5c\" (UID: \"cc019303-87b0-499a-86c1-af977b6cce5c\") " Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.583681 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities" (OuterVolumeSpecName: "utilities") pod "cc019303-87b0-499a-86c1-af977b6cce5c" (UID: "cc019303-87b0-499a-86c1-af977b6cce5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.589353 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79" (OuterVolumeSpecName: "kube-api-access-bwd79") pod "cc019303-87b0-499a-86c1-af977b6cce5c" (UID: "cc019303-87b0-499a-86c1-af977b6cce5c"). InnerVolumeSpecName "kube-api-access-bwd79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.626699 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc019303-87b0-499a-86c1-af977b6cce5c" (UID: "cc019303-87b0-499a-86c1-af977b6cce5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.684709 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwd79\" (UniqueName: \"kubernetes.io/projected/cc019303-87b0-499a-86c1-af977b6cce5c-kube-api-access-bwd79\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.684745 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.684760 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019303-87b0-499a-86c1-af977b6cce5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.873368 4959 generic.go:334] "Generic (PLEG): container finished" podID="cc019303-87b0-499a-86c1-af977b6cce5c" containerID="6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32" exitCode=0 Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.873419 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerDied","Data":"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32"} Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.873451 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg5wt" event={"ID":"cc019303-87b0-499a-86c1-af977b6cce5c","Type":"ContainerDied","Data":"4d26df41022d2f9f673f139ac10f41ee69e7630f318db1fb890da62d43074d12"} Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.873474 4959 scope.go:117] "RemoveContainer" containerID="6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.873475 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg5wt" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.916953 4959 scope.go:117] "RemoveContainer" containerID="df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.926126 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.936054 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zg5wt"] Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.955439 4959 scope.go:117] "RemoveContainer" containerID="7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.985693 4959 scope.go:117] "RemoveContainer" containerID="6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32" Oct 03 16:11:28 crc kubenswrapper[4959]: E1003 16:11:28.986202 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32\": container with ID starting with 6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32 not found: ID does not exist" containerID="6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.986254 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32"} err="failed to get container status \"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32\": rpc error: code = NotFound desc = could not find container \"6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32\": container with ID starting with 6ba3be072f372382bdd904bb8d48b6616151f96f7d7540b07da03719d9dc0b32 not found: ID does not exist" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.986285 4959 scope.go:117] "RemoveContainer" containerID="df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6" Oct 03 16:11:28 crc kubenswrapper[4959]: E1003 16:11:28.987258 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6\": container with ID starting with df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6 not found: ID does not exist" containerID="df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.987310 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6"} err="failed to get container status \"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6\": rpc error: code = NotFound desc = could not find container \"df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6\": container with ID starting with df3272b4ee7ead602b911aaede02b0f8297036223c8ad653cc9d5fd4e8cbebe6 not found: ID does not exist" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.987341 4959 scope.go:117] "RemoveContainer" containerID="7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29" Oct 03 16:11:28 crc kubenswrapper[4959]: E1003 16:11:28.987657 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29\": container with ID starting with 7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29 not found: ID does not exist" containerID="7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29" Oct 03 16:11:28 crc kubenswrapper[4959]: I1003 16:11:28.987698 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29"} err="failed to get container status \"7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29\": rpc error: code = NotFound desc = could not find container \"7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29\": container with ID starting with 7a893286bdea594bdced672e320dadb507b2b8af49eab345563707a27cc93e29 not found: ID does not exist" Oct 03 16:11:29 crc kubenswrapper[4959]: I1003 16:11:29.709334 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" path="/var/lib/kubelet/pods/cc019303-87b0-499a-86c1-af977b6cce5c/volumes" Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.044829 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.045673 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.045728 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.046758 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.046832 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e" gracePeriod=600 Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.979781 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e" exitCode=0 Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.980705 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e"} Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.980755 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437"} Oct 03 16:11:36 crc kubenswrapper[4959]: I1003 16:11:36.980801 4959 scope.go:117] "RemoveContainer" containerID="49d90870ff0c89915fb8c4275fd35fc15714c87e7b3cb332fa93f6fa4356b093" Oct 03 16:13:36 crc kubenswrapper[4959]: I1003 16:13:36.044134 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:13:36 crc kubenswrapper[4959]: I1003 16:13:36.044799 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:14:06 crc kubenswrapper[4959]: I1003 16:14:06.045813 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:14:06 crc kubenswrapper[4959]: I1003 16:14:06.047989 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:14:20 crc kubenswrapper[4959]: I1003 16:14:20.894724 4959 generic.go:334] "Generic (PLEG): container finished" podID="338c9aed-0345-4607-b087-b2523f403bd4" containerID="eeee0271ab54afbb3a094abb28b512877a20c1395697aae5b7f794a9de1acb8e" exitCode=0 Oct 03 16:14:20 crc kubenswrapper[4959]: I1003 16:14:20.894937 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" event={"ID":"338c9aed-0345-4607-b087-b2523f403bd4","Type":"ContainerDied","Data":"eeee0271ab54afbb3a094abb28b512877a20c1395697aae5b7f794a9de1acb8e"} Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.385303 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444440 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444500 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444556 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444572 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm8f7\" (UniqueName: \"kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444649 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444665 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444679 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444764 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444833 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444879 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.444908 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1\") pod \"338c9aed-0345-4607-b087-b2523f403bd4\" (UID: \"338c9aed-0345-4607-b087-b2523f403bd4\") " Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.464510 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph" (OuterVolumeSpecName: "ceph") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.478746 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.490363 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.490708 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7" (OuterVolumeSpecName: "kube-api-access-qm8f7") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "kube-api-access-qm8f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.503053 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.522407 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.522537 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.529595 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.529977 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.533083 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory" (OuterVolumeSpecName: "inventory") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.534900 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "338c9aed-0345-4607-b087-b2523f403bd4" (UID: "338c9aed-0345-4607-b087-b2523f403bd4"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547856 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547911 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547921 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547932 4959 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547941 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm8f7\" (UniqueName: \"kubernetes.io/projected/338c9aed-0345-4607-b087-b2523f403bd4-kube-api-access-qm8f7\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547950 4959 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547959 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547990 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.547998 4959 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.548007 4959 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338c9aed-0345-4607-b087-b2523f403bd4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.548014 4959 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/338c9aed-0345-4607-b087-b2523f403bd4-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.914310 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" event={"ID":"338c9aed-0345-4607-b087-b2523f403bd4","Type":"ContainerDied","Data":"2714a67aafc3668538a8dc8d50d49ce8656343a1a98d8d9e1f7acca6aff508d7"} Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.914370 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2714a67aafc3668538a8dc8d50d49ce8656343a1a98d8d9e1f7acca6aff508d7" Oct 03 16:14:22 crc kubenswrapper[4959]: I1003 16:14:22.914373 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq" Oct 03 16:14:36 crc kubenswrapper[4959]: I1003 16:14:36.044071 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:14:36 crc kubenswrapper[4959]: I1003 16:14:36.045011 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:14:36 crc kubenswrapper[4959]: I1003 16:14:36.045101 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 16:14:36 crc kubenswrapper[4959]: I1003 16:14:36.046726 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:14:36 crc kubenswrapper[4959]: I1003 16:14:36.046980 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" gracePeriod=600 Oct 03 16:14:36 crc kubenswrapper[4959]: E1003 16:14:36.171946 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:14:37 crc kubenswrapper[4959]: I1003 16:14:37.084378 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" exitCode=0 Oct 03 16:14:37 crc kubenswrapper[4959]: I1003 16:14:37.084437 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437"} Oct 03 16:14:37 crc kubenswrapper[4959]: I1003 16:14:37.084470 4959 scope.go:117] "RemoveContainer" containerID="c758657121d0e6ab67c25c96afa6da71533c483a54697d46e15b36bbf1eccb3e" Oct 03 16:14:37 crc kubenswrapper[4959]: I1003 16:14:37.085168 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:14:37 crc kubenswrapper[4959]: E1003 16:14:37.085669 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:14:48 crc kubenswrapper[4959]: I1003 16:14:48.685958 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:14:48 crc kubenswrapper[4959]: E1003 16:14:48.687292 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:14:58 crc kubenswrapper[4959]: E1003 16:14:58.616489 4959 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.245:49360->38.129.56.245:39745: read tcp 38.129.56.245:49360->38.129.56.245:39745: read: connection reset by peer Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.160721 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx"] Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.162422 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338c9aed-0345-4607-b087-b2523f403bd4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.162522 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="338c9aed-0345-4607-b087-b2523f403bd4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.162595 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="extract-content" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.162653 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="extract-content" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.162719 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.162780 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.162860 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.162944 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.163016 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="extract-utilities" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163074 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="extract-utilities" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.163140 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="extract-utilities" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163218 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="extract-utilities" Oct 03 16:15:00 crc kubenswrapper[4959]: E1003 16:15:00.163308 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="extract-content" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163379 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="extract-content" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163675 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="338c9aed-0345-4607-b087-b2523f403bd4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163783 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc019303-87b0-499a-86c1-af977b6cce5c" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.163877 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="59e33877-c720-4324-ae79-515b18cb0693" containerName="registry-server" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.164925 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.172413 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx"] Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.175895 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.176292 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.272841 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.272935 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.273000 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kkvk\" (UniqueName: \"kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.374848 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kkvk\" (UniqueName: \"kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.375243 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.375484 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.376433 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.388742 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.392972 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kkvk\" (UniqueName: \"kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk\") pod \"collect-profiles-29325135-rb6jx\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:00 crc kubenswrapper[4959]: I1003 16:15:00.502486 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:01 crc kubenswrapper[4959]: I1003 16:15:01.000062 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx"] Oct 03 16:15:01 crc kubenswrapper[4959]: W1003 16:15:01.002260 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ec6539e_fc4e_46ef_aa02_935c28e52b85.slice/crio-5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877 WatchSource:0}: Error finding container 5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877: Status 404 returned error can't find the container with id 5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877 Oct 03 16:15:01 crc kubenswrapper[4959]: I1003 16:15:01.406387 4959 generic.go:334] "Generic (PLEG): container finished" podID="0ec6539e-fc4e-46ef-aa02-935c28e52b85" containerID="66138016a988d8c2397bf2f95c6edeae1234e7d41cd8f89374823fbdbbde9fcd" exitCode=0 Oct 03 16:15:01 crc kubenswrapper[4959]: I1003 16:15:01.406429 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" event={"ID":"0ec6539e-fc4e-46ef-aa02-935c28e52b85","Type":"ContainerDied","Data":"66138016a988d8c2397bf2f95c6edeae1234e7d41cd8f89374823fbdbbde9fcd"} Oct 03 16:15:01 crc kubenswrapper[4959]: I1003 16:15:01.406644 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" event={"ID":"0ec6539e-fc4e-46ef-aa02-935c28e52b85","Type":"ContainerStarted","Data":"5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877"} Oct 03 16:15:01 crc kubenswrapper[4959]: I1003 16:15:01.696023 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:15:01 crc kubenswrapper[4959]: E1003 16:15:01.696673 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.811832 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.931398 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kkvk\" (UniqueName: \"kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk\") pod \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.931557 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume\") pod \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.931611 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume\") pod \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\" (UID: \"0ec6539e-fc4e-46ef-aa02-935c28e52b85\") " Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.932802 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume" (OuterVolumeSpecName: "config-volume") pod "0ec6539e-fc4e-46ef-aa02-935c28e52b85" (UID: "0ec6539e-fc4e-46ef-aa02-935c28e52b85"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.938812 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0ec6539e-fc4e-46ef-aa02-935c28e52b85" (UID: "0ec6539e-fc4e-46ef-aa02-935c28e52b85"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:15:02 crc kubenswrapper[4959]: I1003 16:15:02.938819 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk" (OuterVolumeSpecName: "kube-api-access-9kkvk") pod "0ec6539e-fc4e-46ef-aa02-935c28e52b85" (UID: "0ec6539e-fc4e-46ef-aa02-935c28e52b85"). InnerVolumeSpecName "kube-api-access-9kkvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.034600 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kkvk\" (UniqueName: \"kubernetes.io/projected/0ec6539e-fc4e-46ef-aa02-935c28e52b85-kube-api-access-9kkvk\") on node \"crc\" DevicePath \"\"" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.034637 4959 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ec6539e-fc4e-46ef-aa02-935c28e52b85-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.034646 4959 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ec6539e-fc4e-46ef-aa02-935c28e52b85-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.425885 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" event={"ID":"0ec6539e-fc4e-46ef-aa02-935c28e52b85","Type":"ContainerDied","Data":"5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877"} Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.426230 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5107db22a318b3ad5913725e9605d69170f77fcc266e3a2612507b183ba20877" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.425921 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-rb6jx" Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.883398 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7"] Oct 03 16:15:03 crc kubenswrapper[4959]: I1003 16:15:03.892881 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325090-5gzs7"] Oct 03 16:15:05 crc kubenswrapper[4959]: I1003 16:15:05.708595 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08619763-d2c8-4d95-955f-c9baaa8d3bd3" path="/var/lib/kubelet/pods/08619763-d2c8-4d95-955f-c9baaa8d3bd3/volumes" Oct 03 16:15:13 crc kubenswrapper[4959]: I1003 16:15:13.432561 4959 scope.go:117] "RemoveContainer" containerID="290a1b47e9abef7b05a5a4cb3ed4620a82639d0869cac250f7fbc3a8e66a8bf6" Oct 03 16:15:14 crc kubenswrapper[4959]: I1003 16:15:14.685514 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:15:14 crc kubenswrapper[4959]: E1003 16:15:14.686655 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:15:26 crc kubenswrapper[4959]: I1003 16:15:26.685369 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:15:26 crc kubenswrapper[4959]: E1003 16:15:26.686289 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:15:37 crc kubenswrapper[4959]: I1003 16:15:37.686291 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:15:37 crc kubenswrapper[4959]: E1003 16:15:37.687034 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:15:48 crc kubenswrapper[4959]: I1003 16:15:48.686153 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:15:48 crc kubenswrapper[4959]: E1003 16:15:48.687397 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:02 crc kubenswrapper[4959]: I1003 16:16:02.685945 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:16:02 crc kubenswrapper[4959]: E1003 16:16:02.686660 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:14 crc kubenswrapper[4959]: I1003 16:16:14.687335 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:16:14 crc kubenswrapper[4959]: E1003 16:16:14.688334 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:25 crc kubenswrapper[4959]: I1003 16:16:25.689073 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:16:25 crc kubenswrapper[4959]: E1003 16:16:25.690872 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:39 crc kubenswrapper[4959]: I1003 16:16:39.685516 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:16:39 crc kubenswrapper[4959]: E1003 16:16:39.686525 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.961691 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgpcj/must-gather-s4tmh"] Oct 03 16:16:50 crc kubenswrapper[4959]: E1003 16:16:50.964468 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec6539e-fc4e-46ef-aa02-935c28e52b85" containerName="collect-profiles" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.964490 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec6539e-fc4e-46ef-aa02-935c28e52b85" containerName="collect-profiles" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.964827 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec6539e-fc4e-46ef-aa02-935c28e52b85" containerName="collect-profiles" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.966413 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.968255 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgpcj"/"openshift-service-ca.crt" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.968433 4959 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mgpcj"/"default-dockercfg-bcs6t" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.968555 4959 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgpcj"/"kube-root-ca.crt" Oct 03 16:16:50 crc kubenswrapper[4959]: I1003 16:16:50.972004 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgpcj/must-gather-s4tmh"] Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.057229 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8fx9\" (UniqueName: \"kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.057302 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.181041 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8fx9\" (UniqueName: \"kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.181144 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.181685 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.198834 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8fx9\" (UniqueName: \"kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9\") pod \"must-gather-s4tmh\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.334028 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.871112 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgpcj/must-gather-s4tmh"] Oct 03 16:16:51 crc kubenswrapper[4959]: I1003 16:16:51.873731 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:16:52 crc kubenswrapper[4959]: I1003 16:16:52.797350 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" event={"ID":"e94006d0-59a2-4f28-be46-0d03d23c1c9a","Type":"ContainerStarted","Data":"77cdaf39b7fbca36407b75fb0913e75d021dfbeaac2102d9b7d0fbb2b3968209"} Oct 03 16:16:54 crc kubenswrapper[4959]: I1003 16:16:54.685521 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:16:54 crc kubenswrapper[4959]: E1003 16:16:54.686038 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:16:56 crc kubenswrapper[4959]: I1003 16:16:56.837560 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" event={"ID":"e94006d0-59a2-4f28-be46-0d03d23c1c9a","Type":"ContainerStarted","Data":"6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080"} Oct 03 16:16:56 crc kubenswrapper[4959]: I1003 16:16:56.838078 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" event={"ID":"e94006d0-59a2-4f28-be46-0d03d23c1c9a","Type":"ContainerStarted","Data":"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83"} Oct 03 16:16:56 crc kubenswrapper[4959]: I1003 16:16:56.854309 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" podStartSLOduration=2.5762458759999998 podStartE2EDuration="6.85428195s" podCreationTimestamp="2025-10-03 16:16:50 +0000 UTC" firstStartedPulling="2025-10-03 16:16:51.873505915 +0000 UTC m=+9981.076849332" lastFinishedPulling="2025-10-03 16:16:56.151541989 +0000 UTC m=+9985.354885406" observedRunningTime="2025-10-03 16:16:56.850380335 +0000 UTC m=+9986.053723782" watchObservedRunningTime="2025-10-03 16:16:56.85428195 +0000 UTC m=+9986.057625387" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.165096 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-hx56f"] Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.166825 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.327434 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8c2h\" (UniqueName: \"kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.327578 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.429122 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.429400 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.429534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8c2h\" (UniqueName: \"kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.453449 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8c2h\" (UniqueName: \"kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h\") pod \"crc-debug-hx56f\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.486317 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:17:01 crc kubenswrapper[4959]: I1003 16:17:01.909270 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" event={"ID":"1d14f477-e9e6-405b-96f5-3054444f77a2","Type":"ContainerStarted","Data":"5c4792c515c77f9036358acf8d67d53f2a7e89afe1f280c5a0b1c6b58642ced8"} Oct 03 16:17:06 crc kubenswrapper[4959]: I1003 16:17:06.686174 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:17:06 crc kubenswrapper[4959]: E1003 16:17:06.687535 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:17:14 crc kubenswrapper[4959]: I1003 16:17:14.045494 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" event={"ID":"1d14f477-e9e6-405b-96f5-3054444f77a2","Type":"ContainerStarted","Data":"de069483f19a9d8ee9e3f03310d74bdff43c64d4b66f84c5a95445e2585a767c"} Oct 03 16:17:14 crc kubenswrapper[4959]: I1003 16:17:14.065305 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" podStartSLOduration=1.459684425 podStartE2EDuration="13.065288474s" podCreationTimestamp="2025-10-03 16:17:01 +0000 UTC" firstStartedPulling="2025-10-03 16:17:01.548903162 +0000 UTC m=+9990.752246579" lastFinishedPulling="2025-10-03 16:17:13.154507201 +0000 UTC m=+10002.357850628" observedRunningTime="2025-10-03 16:17:14.059901274 +0000 UTC m=+10003.263244701" watchObservedRunningTime="2025-10-03 16:17:14.065288474 +0000 UTC m=+10003.268631881" Oct 03 16:17:17 crc kubenswrapper[4959]: I1003 16:17:17.690127 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:17:17 crc kubenswrapper[4959]: E1003 16:17:17.690988 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:17:28 crc kubenswrapper[4959]: I1003 16:17:28.685804 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:17:28 crc kubenswrapper[4959]: E1003 16:17:28.686402 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:17:43 crc kubenswrapper[4959]: I1003 16:17:43.685347 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:17:43 crc kubenswrapper[4959]: E1003 16:17:43.685930 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:17:54 crc kubenswrapper[4959]: I1003 16:17:54.686536 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:17:54 crc kubenswrapper[4959]: E1003 16:17:54.687275 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:18:09 crc kubenswrapper[4959]: I1003 16:18:09.687784 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:18:09 crc kubenswrapper[4959]: E1003 16:18:09.688501 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:18:24 crc kubenswrapper[4959]: I1003 16:18:24.686169 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:18:24 crc kubenswrapper[4959]: E1003 16:18:24.687005 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:18:31 crc kubenswrapper[4959]: I1003 16:18:31.460573 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_17939c3e-ae90-4f2f-983b-7f4011b95ef2/init-config-reloader/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.088904 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_17939c3e-ae90-4f2f-983b-7f4011b95ef2/init-config-reloader/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.095963 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_17939c3e-ae90-4f2f-983b-7f4011b95ef2/alertmanager/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.269244 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_17939c3e-ae90-4f2f-983b-7f4011b95ef2/config-reloader/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.463064 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_614f87cc-6308-4228-aad1-a0eafb9f57a8/aodh-api/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.544653 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_614f87cc-6308-4228-aad1-a0eafb9f57a8/aodh-evaluator/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.703840 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_614f87cc-6308-4228-aad1-a0eafb9f57a8/aodh-listener/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.716252 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_614f87cc-6308-4228-aad1-a0eafb9f57a8/aodh-notifier/0.log" Oct 03 16:18:32 crc kubenswrapper[4959]: I1003 16:18:32.904900 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f4b8ff658-pvrk8_e85cb5d9-0483-4ebf-979f-c614249eaa4d/barbican-api/0.log" Oct 03 16:18:33 crc kubenswrapper[4959]: I1003 16:18:33.082556 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7f4b8ff658-pvrk8_e85cb5d9-0483-4ebf-979f-c614249eaa4d/barbican-api-log/0.log" Oct 03 16:18:33 crc kubenswrapper[4959]: I1003 16:18:33.754015 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d6558856b-h9zfg_217643a2-9b13-4460-b9bd-7e01a2235174/barbican-keystone-listener/0.log" Oct 03 16:18:33 crc kubenswrapper[4959]: I1003 16:18:33.850859 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6d6558856b-h9zfg_217643a2-9b13-4460-b9bd-7e01a2235174/barbican-keystone-listener-log/0.log" Oct 03 16:18:33 crc kubenswrapper[4959]: I1003 16:18:33.944488 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d48b6b55f-s88sc_e10ff4ef-9f81-416c-a313-91c8313d2104/barbican-worker/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.024043 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d48b6b55f-s88sc_e10ff4ef-9f81-416c-a313-91c8313d2104/barbican-worker-log/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.142460 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-6vxnz_faa34915-61b3-4e8f-a90b-b4d45cace539/bootstrap-openstack-openstack-cell1/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.300886 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/ceilometer-central-agent/1.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.346639 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/ceilometer-central-agent/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.497439 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/ceilometer-notification-agent/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.533495 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/proxy-httpd/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.539266 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/ceilometer-notification-agent/1.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.697770 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5922a823-0599-4108-a7ca-d7eb68de6c2d/sg-core/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.726989 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-xp2jr_b1a9268b-3fd9-4a2c-95b9-df1c8e67ea88/ceph-client-openstack-openstack-cell1/0.log" Oct 03 16:18:34 crc kubenswrapper[4959]: I1003 16:18:34.970199 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae/cinder-api-log/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.013296 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b5e1778c-6d55-4e02-9b2e-d7ce9f9ebeae/cinder-api/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.236871 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c4d9369a-f74d-4c94-8578-41c4cd5beb0d/probe/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.311832 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c4d9369a-f74d-4c94-8578-41c4cd5beb0d/cinder-backup/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.451524 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6355c44f-1794-4bb7-9c2d-f71963345c81/cinder-scheduler/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.543161 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6355c44f-1794-4bb7-9c2d-f71963345c81/probe/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.681954 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_363bf001-9d71-4656-ab22-9bc29fbb88c3/cinder-volume/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.764100 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_363bf001-9d71-4656-ab22-9bc29fbb88c3/probe/0.log" Oct 03 16:18:35 crc kubenswrapper[4959]: I1003 16:18:35.947009 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-vm92f_8d721f05-6907-49fa-87ed-de92d1fa8139/configure-network-openstack-openstack-cell1/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.103878 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-hnc2k_b3b15986-a162-4643-a76e-67d33d624642/configure-os-openstack-openstack-cell1/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.214078 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d6cd869d9-km27m_30fd93eb-4baf-471e-9dd3-28a623a2a954/init/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.458527 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d6cd869d9-km27m_30fd93eb-4baf-471e-9dd3-28a623a2a954/dnsmasq-dns/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.476412 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d6cd869d9-km27m_30fd93eb-4baf-471e-9dd3-28a623a2a954/init/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.667589 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-jlh5w_7b050197-797a-4c1d-8e40-530801ae7bd3/download-cache-openstack-openstack-cell1/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.685693 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:18:36 crc kubenswrapper[4959]: E1003 16:18:36.686746 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.752796 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7/glance-httpd/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.857695 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9c3c7bd2-ca02-41c3-8b91-cd1ba06a81f7/glance-log/0.log" Oct 03 16:18:36 crc kubenswrapper[4959]: I1003 16:18:36.954164 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03/glance-httpd/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.090495 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c5185f5a-c1ec-4beb-9b1d-79f01f3a0d03/glance-log/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.383093 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-799957bf96-8r76b_e0417f0e-e43d-4ade-83f6-189961bdf2a9/heat-api/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.540021 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7fc99996f4-zzpdp_5fcd2bf1-cf03-4438-930d-d87923613b4c/heat-cfnapi/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.617214 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-67486f58d7-k7xmb_7aa03178-4f69-410a-9a12-62bdbf7c556d/heat-engine/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.898989 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5656c49bff-fgjq5_8ce9411f-dcdd-4538-a83e-193382cda8e3/horizon-log/0.log" Oct 03 16:18:37 crc kubenswrapper[4959]: I1003 16:18:37.925633 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5656c49bff-fgjq5_8ce9411f-dcdd-4538-a83e-193382cda8e3/horizon/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.143780 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-w7ksg_46ba8fc5-d9cf-4e16-aba1-b9145d389702/install-certs-openstack-openstack-cell1/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.259086 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-b85rz_f3102942-6ad8-45bf-a609-90bdd43c2176/install-os-openstack-openstack-cell1/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.618670 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325061-jxq5l_7403bf5f-ba18-42cb-8d44-f5ed5d463efd/keystone-cron/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.625602 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8fdb48fc-6jdsh_47e5037e-65ea-42d7-aa93-9c7d9d2e4a2f/keystone-api/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.803854 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325121-h72rz_3a6a8927-b3b3-4aaf-8f6a-b9a5daff2c13/keystone-cron/0.log" Oct 03 16:18:38 crc kubenswrapper[4959]: I1003 16:18:38.991130 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8ffac68e-7637-40b7-a7fd-951926faac16/kube-state-metrics/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.070045 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-nxgb8_86133afd-9abd-4209-bdc7-b729e26d8766/libvirt-openstack-openstack-cell1/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.279068 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_a804906c-740a-4406-bdd5-44ab84050ee8/manila-api-log/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.292561 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_a804906c-740a-4406-bdd5-44ab84050ee8/manila-api/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.697782 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910/probe/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.706899 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_ddcfebfd-3f2e-4ad3-a1d1-192b4f51b910/manila-scheduler/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.944365 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_2e142326-f734-4066-9b6c-d7a587851ca5/probe/0.log" Oct 03 16:18:39 crc kubenswrapper[4959]: I1003 16:18:39.955909 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_2e142326-f734-4066-9b6c-d7a587851ca5/manila-share/0.log" Oct 03 16:18:40 crc kubenswrapper[4959]: I1003 16:18:40.134636 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-copy-data_19472efc-08e5-4271-ada9-a7664d2dcef6/adoption/0.log" Oct 03 16:18:40 crc kubenswrapper[4959]: I1003 16:18:40.571841 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6f668759cc-jmsxf_e4389945-7007-4c05-b044-2de169b96b46/neutron-api/0.log" Oct 03 16:18:40 crc kubenswrapper[4959]: I1003 16:18:40.665144 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6f668759cc-jmsxf_e4389945-7007-4c05-b044-2de169b96b46/neutron-httpd/0.log" Oct 03 16:18:40 crc kubenswrapper[4959]: I1003 16:18:40.926687 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-2c2rq_2eba6fae-31eb-4f5f-9d23-68f490fe5fa4/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 03 16:18:41 crc kubenswrapper[4959]: I1003 16:18:41.224401 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-877z2_5f58accc-ca6d-4db4-b69c-200fb1c1e35e/neutron-metadata-openstack-openstack-cell1/0.log" Oct 03 16:18:42 crc kubenswrapper[4959]: I1003 16:18:42.154635 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-chj9r_fbd008c2-83fd-4c30-a3e6-6eb0e52965f7/neutron-sriov-openstack-openstack-cell1/0.log" Oct 03 16:18:42 crc kubenswrapper[4959]: I1003 16:18:42.415479 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e2e48614-e460-4107-8548-78d3720299b7/nova-api-api/0.log" Oct 03 16:18:42 crc kubenswrapper[4959]: I1003 16:18:42.503497 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e2e48614-e460-4107-8548-78d3720299b7/nova-api-log/0.log" Oct 03 16:18:42 crc kubenswrapper[4959]: I1003 16:18:42.791298 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f24ab41d-8257-4db6-8c24-44a157bfa8b7/nova-cell0-conductor-conductor/0.log" Oct 03 16:18:43 crc kubenswrapper[4959]: I1003 16:18:43.092489 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4e9434fc-ae1c-42e7-95de-3735c19e0c1f/nova-cell1-conductor-conductor/0.log" Oct 03 16:18:43 crc kubenswrapper[4959]: I1003 16:18:43.249162 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_57e3c964-7bc5-41a1-9c77-0f42cffb4628/memcached/0.log" Oct 03 16:18:43 crc kubenswrapper[4959]: I1003 16:18:43.886012 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_84757a0b-ee0a-443c-8b44-b3aeaae6785e/nova-cell1-novncproxy-novncproxy/0.log" Oct 03 16:18:44 crc kubenswrapper[4959]: I1003 16:18:44.054802 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellh9jmq_338c9aed-0345-4607-b087-b2523f403bd4/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 03 16:18:44 crc kubenswrapper[4959]: I1003 16:18:44.456639 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-8w98r_2a09699c-f954-45ac-b96b-accef4753ab3/nova-cell1-openstack-openstack-cell1/0.log" Oct 03 16:18:44 crc kubenswrapper[4959]: I1003 16:18:44.596783 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f2b08809-1a97-4f9c-835f-0f794d9adf09/nova-metadata-log/0.log" Oct 03 16:18:44 crc kubenswrapper[4959]: I1003 16:18:44.737410 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f2b08809-1a97-4f9c-835f-0f794d9adf09/nova-metadata-metadata/0.log" Oct 03 16:18:44 crc kubenswrapper[4959]: I1003 16:18:44.966349 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a18341b8-2b61-438d-afe0-8598f362b247/nova-scheduler-scheduler/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.074691 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5664bbcc7d-snqwd_b0c747c0-dca4-4c05-a8ee-9b68b6e2797a/init/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.239881 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5664bbcc7d-snqwd_b0c747c0-dca4-4c05-a8ee-9b68b6e2797a/init/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.277799 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5664bbcc7d-snqwd_b0c747c0-dca4-4c05-a8ee-9b68b6e2797a/octavia-api-provider-agent/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.407965 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-5664bbcc7d-snqwd_b0c747c0-dca4-4c05-a8ee-9b68b6e2797a/octavia-api/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.505343 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-lqvnf_88d6d698-a05d-4d67-8844-bef19cb4689e/init/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.737545 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-lqvnf_88d6d698-a05d-4d67-8844-bef19cb4689e/init/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.792973 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-lqvnf_88d6d698-a05d-4d67-8844-bef19cb4689e/octavia-healthmanager/0.log" Oct 03 16:18:45 crc kubenswrapper[4959]: I1003 16:18:45.958864 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zqfhm_6b0768c0-0cb6-423e-af94-dd2be48aa3a4/init/0.log" Oct 03 16:18:46 crc kubenswrapper[4959]: I1003 16:18:46.226627 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zqfhm_6b0768c0-0cb6-423e-af94-dd2be48aa3a4/init/0.log" Oct 03 16:18:46 crc kubenswrapper[4959]: I1003 16:18:46.245038 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-zqfhm_6b0768c0-0cb6-423e-af94-dd2be48aa3a4/octavia-housekeeping/0.log" Oct 03 16:18:46 crc kubenswrapper[4959]: I1003 16:18:46.407295 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-zt5s2_b24d93bb-3ffd-43c6-9c19-377d84d71cde/init/0.log" Oct 03 16:18:46 crc kubenswrapper[4959]: I1003 16:18:46.803525 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-zt5s2_b24d93bb-3ffd-43c6-9c19-377d84d71cde/octavia-amphora-httpd/0.log" Oct 03 16:18:46 crc kubenswrapper[4959]: I1003 16:18:46.861903 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-zt5s2_b24d93bb-3ffd-43c6-9c19-377d84d71cde/init/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.100443 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-n5vz4_92238071-8adc-41d2-be43-7b92ba7565c7/init/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.344257 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-n5vz4_92238071-8adc-41d2-be43-7b92ba7565c7/init/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.383265 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-n5vz4_92238071-8adc-41d2-be43-7b92ba7565c7/octavia-rsyslog/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.541303 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-nc45h_26e1c5e1-5763-4b18-8410-19f956516874/init/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.719736 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-nc45h_26e1c5e1-5763-4b18-8410-19f956516874/init/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.838566 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-nc45h_26e1c5e1-5763-4b18-8410-19f956516874/octavia-worker/0.log" Oct 03 16:18:47 crc kubenswrapper[4959]: I1003 16:18:47.964422 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_de510575-5126-4c3a-9c24-15406b979fcf/mysql-bootstrap/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.151326 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_de510575-5126-4c3a-9c24-15406b979fcf/galera/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.157403 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_de510575-5126-4c3a-9c24-15406b979fcf/mysql-bootstrap/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.346850 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b654062a-f888-48b8-8b60-4fb64546dfe0/mysql-bootstrap/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.559227 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b654062a-f888-48b8-8b60-4fb64546dfe0/mysql-bootstrap/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.571057 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b654062a-f888-48b8-8b60-4fb64546dfe0/galera/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.685842 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:18:48 crc kubenswrapper[4959]: E1003 16:18:48.686132 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.768813 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_236209a8-8d65-4c3f-9803-824fca99dd36/openstackclient/0.log" Oct 03 16:18:48 crc kubenswrapper[4959]: I1003 16:18:48.874743 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8nphv_973e0edf-ad3a-4987-a8e6-4724186d830d/ovn-controller/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.048156 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-zpqwq_6269efac-eb85-49fe-b427-4b5cc0d50fdd/openstack-network-exporter/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.247606 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p6zgv_9b1dcf14-6652-4948-a200-d533be359fd0/ovsdb-server-init/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.449718 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p6zgv_9b1dcf14-6652-4948-a200-d533be359fd0/ovsdb-server/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.468953 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p6zgv_9b1dcf14-6652-4948-a200-d533be359fd0/ovsdb-server-init/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.472846 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-p6zgv_9b1dcf14-6652-4948-a200-d533be359fd0/ovs-vswitchd/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.639749 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-copy-data_5d211fb6-eef0-4a11-9dcf-9cdd00943b05/adoption/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.831569 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f7263f47-e4e8-4b64-b043-45d3dfc60c2f/openstack-network-exporter/0.log" Oct 03 16:18:49 crc kubenswrapper[4959]: I1003 16:18:49.933864 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f7263f47-e4e8-4b64-b043-45d3dfc60c2f/ovn-northd/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.088683 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-scpcf_8df4e8ab-06f9-4a37-8e51-5954f63f2cf3/ovn-openstack-openstack-cell1/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.231851 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5/openstack-network-exporter/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.340465 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b9d1ad77-fd11-4bc2-8d7d-dec4f6652df5/ovsdbserver-nb/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.466339 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_1ad50a25-04e2-433b-af7d-1d73ddf151a9/openstack-network-exporter/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.696098 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_1ad50a25-04e2-433b-af7d-1d73ddf151a9/ovsdbserver-nb/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.849915 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_76dc2579-070a-44a4-8063-89979bc14f52/openstack-network-exporter/0.log" Oct 03 16:18:50 crc kubenswrapper[4959]: I1003 16:18:50.979392 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_76dc2579-070a-44a4-8063-89979bc14f52/ovsdbserver-nb/0.log" Oct 03 16:18:51 crc kubenswrapper[4959]: I1003 16:18:51.106647 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97c621eb-b91e-42fb-91df-25cb1782d264/openstack-network-exporter/0.log" Oct 03 16:18:51 crc kubenswrapper[4959]: I1003 16:18:51.432977 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_97c621eb-b91e-42fb-91df-25cb1782d264/ovsdbserver-sb/0.log" Oct 03 16:18:51 crc kubenswrapper[4959]: I1003 16:18:51.527154 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_053f4c08-15f4-4adb-8b32-e3e200fd4197/openstack-network-exporter/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.262574 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_053f4c08-15f4-4adb-8b32-e3e200fd4197/ovsdbserver-sb/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.318750 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_29b1ce1a-4945-46ba-994b-4cf3167e2070/openstack-network-exporter/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.446927 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_29b1ce1a-4945-46ba-994b-4cf3167e2070/ovsdbserver-sb/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.618901 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-544547cb8d-86sh8_bff4d98d-ab89-42e4-b79c-c7088858d882/placement-api/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.705528 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-544547cb8d-86sh8_bff4d98d-ab89-42e4-b79c-c7088858d882/placement-log/0.log" Oct 03 16:18:52 crc kubenswrapper[4959]: I1003 16:18:52.834125 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-ctz6s6_21dadec9-3ba5-4010-a9ea-e65cbe8cbca3/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 03 16:18:53 crc kubenswrapper[4959]: I1003 16:18:53.035490 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8fcdfb-38cd-4922-91e9-b0a3be243d78/init-config-reloader/0.log" Oct 03 16:18:53 crc kubenswrapper[4959]: I1003 16:18:53.226140 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8fcdfb-38cd-4922-91e9-b0a3be243d78/config-reloader/0.log" Oct 03 16:18:53 crc kubenswrapper[4959]: I1003 16:18:53.233290 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8fcdfb-38cd-4922-91e9-b0a3be243d78/init-config-reloader/0.log" Oct 03 16:18:53 crc kubenswrapper[4959]: I1003 16:18:53.836695 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8fcdfb-38cd-4922-91e9-b0a3be243d78/prometheus/0.log" Oct 03 16:18:53 crc kubenswrapper[4959]: I1003 16:18:53.867447 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_fb8fcdfb-38cd-4922-91e9-b0a3be243d78/thanos-sidecar/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.040075 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ac2cbd4-9c05-48d2-9c80-fad3bb420edc/setup-container/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.242175 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ac2cbd4-9c05-48d2-9c80-fad3bb420edc/setup-container/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.292859 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ac2cbd4-9c05-48d2-9c80-fad3bb420edc/rabbitmq/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.447885 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc4d11dc-89c5-49d3-a63b-44ef60b0489c/setup-container/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.684273 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc4d11dc-89c5-49d3-a63b-44ef60b0489c/setup-container/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.907045 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-mtg8s_4a88799d-3ffb-40b1-b444-0300adede7bc/reboot-os-openstack-openstack-cell1/0.log" Oct 03 16:18:54 crc kubenswrapper[4959]: I1003 16:18:54.936991 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc4d11dc-89c5-49d3-a63b-44ef60b0489c/rabbitmq/0.log" Oct 03 16:18:55 crc kubenswrapper[4959]: I1003 16:18:55.115048 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-c9lbc_f37d30c9-da2b-4894-9f1f-30c37f9b0b19/run-os-openstack-openstack-cell1/0.log" Oct 03 16:18:55 crc kubenswrapper[4959]: I1003 16:18:55.227786 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-j9ckb_521fed84-00a8-449a-a036-dda24250ddc4/ssh-known-hosts-openstack/0.log" Oct 03 16:18:55 crc kubenswrapper[4959]: I1003 16:18:55.416537 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-h4zpc_ca1f3663-0568-4600-9c69-8a92589523ca/telemetry-openstack-openstack-cell1/0.log" Oct 03 16:18:55 crc kubenswrapper[4959]: I1003 16:18:55.589781 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-rt52g_034c6ab7-82c8-4922-ae56-e1862f88d39e/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 03 16:18:55 crc kubenswrapper[4959]: I1003 16:18:55.746864 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-kw7hh_33563547-3b0f-4647-8aff-1a068907aa2d/validate-network-openstack-openstack-cell1/0.log" Oct 03 16:19:01 crc kubenswrapper[4959]: I1003 16:19:01.692818 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:19:01 crc kubenswrapper[4959]: E1003 16:19:01.693436 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:19:14 crc kubenswrapper[4959]: I1003 16:19:14.686341 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:19:14 crc kubenswrapper[4959]: E1003 16:19:14.687135 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:19:26 crc kubenswrapper[4959]: I1003 16:19:26.686373 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:19:26 crc kubenswrapper[4959]: E1003 16:19:26.689831 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:19:29 crc kubenswrapper[4959]: I1003 16:19:29.423895 4959 generic.go:334] "Generic (PLEG): container finished" podID="1d14f477-e9e6-405b-96f5-3054444f77a2" containerID="de069483f19a9d8ee9e3f03310d74bdff43c64d4b66f84c5a95445e2585a767c" exitCode=0 Oct 03 16:19:29 crc kubenswrapper[4959]: I1003 16:19:29.424011 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" event={"ID":"1d14f477-e9e6-405b-96f5-3054444f77a2","Type":"ContainerDied","Data":"de069483f19a9d8ee9e3f03310d74bdff43c64d4b66f84c5a95445e2585a767c"} Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.582092 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.632974 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-hx56f"] Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.643330 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-hx56f"] Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.731468 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8c2h\" (UniqueName: \"kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h\") pod \"1d14f477-e9e6-405b-96f5-3054444f77a2\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.731827 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host\") pod \"1d14f477-e9e6-405b-96f5-3054444f77a2\" (UID: \"1d14f477-e9e6-405b-96f5-3054444f77a2\") " Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.731919 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host" (OuterVolumeSpecName: "host") pod "1d14f477-e9e6-405b-96f5-3054444f77a2" (UID: "1d14f477-e9e6-405b-96f5-3054444f77a2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.732732 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1d14f477-e9e6-405b-96f5-3054444f77a2-host\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.738441 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h" (OuterVolumeSpecName: "kube-api-access-s8c2h") pod "1d14f477-e9e6-405b-96f5-3054444f77a2" (UID: "1d14f477-e9e6-405b-96f5-3054444f77a2"). InnerVolumeSpecName "kube-api-access-s8c2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:19:30 crc kubenswrapper[4959]: I1003 16:19:30.836437 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8c2h\" (UniqueName: \"kubernetes.io/projected/1d14f477-e9e6-405b-96f5-3054444f77a2-kube-api-access-s8c2h\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.450713 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c4792c515c77f9036358acf8d67d53f2a7e89afe1f280c5a0b1c6b58642ced8" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.450787 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-hx56f" Oct 03 16:19:31 crc kubenswrapper[4959]: E1003 16:19:31.692848 4959 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d14f477_e9e6_405b_96f5_3054444f77a2.slice/crio-5c4792c515c77f9036358acf8d67d53f2a7e89afe1f280c5a0b1c6b58642ced8\": RecentStats: unable to find data in memory cache]" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.719532 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d14f477-e9e6-405b-96f5-3054444f77a2" path="/var/lib/kubelet/pods/1d14f477-e9e6-405b-96f5-3054444f77a2/volumes" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.841117 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-lj49w"] Oct 03 16:19:31 crc kubenswrapper[4959]: E1003 16:19:31.841650 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d14f477-e9e6-405b-96f5-3054444f77a2" containerName="container-00" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.841673 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d14f477-e9e6-405b-96f5-3054444f77a2" containerName="container-00" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.841953 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d14f477-e9e6-405b-96f5-3054444f77a2" containerName="container-00" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.843016 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.963241 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5k87\" (UniqueName: \"kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:31 crc kubenswrapper[4959]: I1003 16:19:31.963613 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.065475 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.065581 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5k87\" (UniqueName: \"kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.065858 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.084796 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5k87\" (UniqueName: \"kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87\") pod \"crc-debug-lj49w\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.163680 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:32 crc kubenswrapper[4959]: I1003 16:19:32.461009 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" event={"ID":"bb29a54a-44ca-4d00-b8f6-b25f9bd94072","Type":"ContainerStarted","Data":"166058a3ababef3556592e4190cd3947ea5a24cecfede1242aab64b17f89c753"} Oct 03 16:19:33 crc kubenswrapper[4959]: I1003 16:19:33.477465 4959 generic.go:334] "Generic (PLEG): container finished" podID="bb29a54a-44ca-4d00-b8f6-b25f9bd94072" containerID="0958c4b64c1ac055d624c6b065b3bddbc3c2257fca5562de5ad83e45e5a80365" exitCode=0 Oct 03 16:19:33 crc kubenswrapper[4959]: I1003 16:19:33.477630 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" event={"ID":"bb29a54a-44ca-4d00-b8f6-b25f9bd94072","Type":"ContainerDied","Data":"0958c4b64c1ac055d624c6b065b3bddbc3c2257fca5562de5ad83e45e5a80365"} Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.604678 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.720013 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5k87\" (UniqueName: \"kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87\") pod \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.720419 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host\") pod \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\" (UID: \"bb29a54a-44ca-4d00-b8f6-b25f9bd94072\") " Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.721072 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host" (OuterVolumeSpecName: "host") pod "bb29a54a-44ca-4d00-b8f6-b25f9bd94072" (UID: "bb29a54a-44ca-4d00-b8f6-b25f9bd94072"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.725929 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87" (OuterVolumeSpecName: "kube-api-access-l5k87") pod "bb29a54a-44ca-4d00-b8f6-b25f9bd94072" (UID: "bb29a54a-44ca-4d00-b8f6-b25f9bd94072"). InnerVolumeSpecName "kube-api-access-l5k87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.822435 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5k87\" (UniqueName: \"kubernetes.io/projected/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-kube-api-access-l5k87\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:34 crc kubenswrapper[4959]: I1003 16:19:34.822467 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bb29a54a-44ca-4d00-b8f6-b25f9bd94072-host\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:35 crc kubenswrapper[4959]: I1003 16:19:35.501928 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" event={"ID":"bb29a54a-44ca-4d00-b8f6-b25f9bd94072","Type":"ContainerDied","Data":"166058a3ababef3556592e4190cd3947ea5a24cecfede1242aab64b17f89c753"} Oct 03 16:19:35 crc kubenswrapper[4959]: I1003 16:19:35.501973 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="166058a3ababef3556592e4190cd3947ea5a24cecfede1242aab64b17f89c753" Oct 03 16:19:35 crc kubenswrapper[4959]: I1003 16:19:35.501994 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-lj49w" Oct 03 16:19:40 crc kubenswrapper[4959]: I1003 16:19:40.685388 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:19:41 crc kubenswrapper[4959]: I1003 16:19:41.567625 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830"} Oct 03 16:19:44 crc kubenswrapper[4959]: I1003 16:19:44.436508 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-lj49w"] Oct 03 16:19:44 crc kubenswrapper[4959]: I1003 16:19:44.461850 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-lj49w"] Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.650337 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-zmvqw"] Oct 03 16:19:45 crc kubenswrapper[4959]: E1003 16:19:45.652370 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb29a54a-44ca-4d00-b8f6-b25f9bd94072" containerName="container-00" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.652499 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb29a54a-44ca-4d00-b8f6-b25f9bd94072" containerName="container-00" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.652900 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb29a54a-44ca-4d00-b8f6-b25f9bd94072" containerName="container-00" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.653937 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.698071 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb29a54a-44ca-4d00-b8f6-b25f9bd94072" path="/var/lib/kubelet/pods/bb29a54a-44ca-4d00-b8f6-b25f9bd94072/volumes" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.815332 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x9s8\" (UniqueName: \"kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.815422 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.918116 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x9s8\" (UniqueName: \"kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.918227 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.918447 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.947317 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x9s8\" (UniqueName: \"kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8\") pod \"crc-debug-zmvqw\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:45 crc kubenswrapper[4959]: I1003 16:19:45.978044 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:46 crc kubenswrapper[4959]: W1003 16:19:46.077452 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40389a9b_dd8d_46d4_8542_75765260e583.slice/crio-e2f411a7484fbb1fcd867aa57c182fb91f2a34e90667d47e024a9c27ac8d07de WatchSource:0}: Error finding container e2f411a7484fbb1fcd867aa57c182fb91f2a34e90667d47e024a9c27ac8d07de: Status 404 returned error can't find the container with id e2f411a7484fbb1fcd867aa57c182fb91f2a34e90667d47e024a9c27ac8d07de Oct 03 16:19:46 crc kubenswrapper[4959]: I1003 16:19:46.631142 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" event={"ID":"40389a9b-dd8d-46d4-8542-75765260e583","Type":"ContainerStarted","Data":"9b1c7174e1974b04749152f56ddae938588d30a1034678fae631df33c6f170c6"} Oct 03 16:19:46 crc kubenswrapper[4959]: I1003 16:19:46.631704 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" event={"ID":"40389a9b-dd8d-46d4-8542-75765260e583","Type":"ContainerStarted","Data":"e2f411a7484fbb1fcd867aa57c182fb91f2a34e90667d47e024a9c27ac8d07de"} Oct 03 16:19:46 crc kubenswrapper[4959]: I1003 16:19:46.683464 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-zmvqw"] Oct 03 16:19:46 crc kubenswrapper[4959]: I1003 16:19:46.698483 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgpcj/crc-debug-zmvqw"] Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.643662 4959 generic.go:334] "Generic (PLEG): container finished" podID="40389a9b-dd8d-46d4-8542-75765260e583" containerID="9b1c7174e1974b04749152f56ddae938588d30a1034678fae631df33c6f170c6" exitCode=0 Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.760754 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.867265 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host\") pod \"40389a9b-dd8d-46d4-8542-75765260e583\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.867375 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host" (OuterVolumeSpecName: "host") pod "40389a9b-dd8d-46d4-8542-75765260e583" (UID: "40389a9b-dd8d-46d4-8542-75765260e583"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.867472 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x9s8\" (UniqueName: \"kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8\") pod \"40389a9b-dd8d-46d4-8542-75765260e583\" (UID: \"40389a9b-dd8d-46d4-8542-75765260e583\") " Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.868062 4959 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40389a9b-dd8d-46d4-8542-75765260e583-host\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.888850 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8" (OuterVolumeSpecName: "kube-api-access-5x9s8") pod "40389a9b-dd8d-46d4-8542-75765260e583" (UID: "40389a9b-dd8d-46d4-8542-75765260e583"). InnerVolumeSpecName "kube-api-access-5x9s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:19:47 crc kubenswrapper[4959]: I1003 16:19:47.970080 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x9s8\" (UniqueName: \"kubernetes.io/projected/40389a9b-dd8d-46d4-8542-75765260e583-kube-api-access-5x9s8\") on node \"crc\" DevicePath \"\"" Oct 03 16:19:48 crc kubenswrapper[4959]: I1003 16:19:48.657069 4959 scope.go:117] "RemoveContainer" containerID="9b1c7174e1974b04749152f56ddae938588d30a1034678fae631df33c6f170c6" Oct 03 16:19:48 crc kubenswrapper[4959]: I1003 16:19:48.657090 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/crc-debug-zmvqw" Oct 03 16:19:49 crc kubenswrapper[4959]: I1003 16:19:49.710728 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40389a9b-dd8d-46d4-8542-75765260e583" path="/var/lib/kubelet/pods/40389a9b-dd8d-46d4-8542-75765260e583/volumes" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.046946 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/util/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.167879 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/util/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.198876 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/pull/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.261502 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/pull/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.423751 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/util/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.427970 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/pull/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.430335 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4bc19491273fb05f10b8b1261ee07db0d4b5cae179a5dad352c99ff354m4m4z_ea1ffdb4-f9b4-42aa-b757-f04040023dc6/extract/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.594090 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-n78z6_1848147c-d228-4d3c-bace-7628a688438d/kube-rbac-proxy/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.708216 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-dsq2k_b991eb57-3139-462c-921b-72525e281d25/kube-rbac-proxy/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.717183 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-n78z6_1848147c-d228-4d3c-bace-7628a688438d/manager/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.895456 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-dsq2k_b991eb57-3139-462c-921b-72525e281d25/manager/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.945722 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-5bgvs_f1efc6bb-e4c4-437f-8a29-8b587327652b/kube-rbac-proxy/0.log" Oct 03 16:20:29 crc kubenswrapper[4959]: I1003 16:20:29.967864 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-5bgvs_f1efc6bb-e4c4-437f-8a29-8b587327652b/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.084293 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-zpl6z_c5519a95-6c42-4bc5-be67-626e8878421c/kube-rbac-proxy/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.239619 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-zpl6z_c5519a95-6c42-4bc5-be67-626e8878421c/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.329845 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fdhll_84bc4832-d055-4c00-955d-72e589902ea7/kube-rbac-proxy/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.387429 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fdhll_84bc4832-d055-4c00-955d-72e589902ea7/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.418250 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-crg46_a1ef21c1-ce09-47cc-b1c6-a7e8916431c6/kube-rbac-proxy/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.515580 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-crg46_a1ef21c1-ce09-47cc-b1c6-a7e8916431c6/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.605063 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-cg44n_e06190ac-1517-453b-a038-8437d0c77ee1/kube-rbac-proxy/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.841454 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-cg44n_e06190ac-1517-453b-a038-8437d0c77ee1/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.880406 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-9ptc9_111756cf-9060-4c4b-878b-4ac998f89786/manager/0.log" Oct 03 16:20:30 crc kubenswrapper[4959]: I1003 16:20:30.901453 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-9ptc9_111756cf-9060-4c4b-878b-4ac998f89786/kube-rbac-proxy/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.060917 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-w96jd_37c195cb-ee19-4040-ae6a-388685fed18d/kube-rbac-proxy/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.206309 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-w96jd_37c195cb-ee19-4040-ae6a-388685fed18d/manager/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.233167 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-55vd5_0f49a60f-299a-4568-81ca-687eb8d439d3/kube-rbac-proxy/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.305090 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-55vd5_0f49a60f-299a-4568-81ca-687eb8d439d3/manager/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.430695 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-qkmbn_13a6dcff-d726-469a-9807-0ff2d76a7713/kube-rbac-proxy/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.457180 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-qkmbn_13a6dcff-d726-469a-9807-0ff2d76a7713/manager/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.519190 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-h29xd_6aa2a2f3-7729-43a4-8564-c74ef4ddf86d/kube-rbac-proxy/0.log" Oct 03 16:20:31 crc kubenswrapper[4959]: I1003 16:20:31.682264 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-h29xd_6aa2a2f3-7729-43a4-8564-c74ef4ddf86d/manager/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.171349 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-ssfjs_469d9704-d797-4ab2-9e51-b4a2a62942b7/kube-rbac-proxy/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.376463 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-ssfjs_469d9704-d797-4ab2-9e51-b4a2a62942b7/manager/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.385508 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-rfvdf_ce1839e2-3240-40dd-9145-867852c68de2/kube-rbac-proxy/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.410832 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678hr24c_f8d0c10b-38a4-435e-a620-6e69d1a31d5c/kube-rbac-proxy/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.420408 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-rfvdf_ce1839e2-3240-40dd-9145-867852c68de2/manager/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.569824 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fd589ff7-nf5sw_f3e29379-18ac-4401-be88-623abc26373e/kube-rbac-proxy/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.593424 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678hr24c_f8d0c10b-38a4-435e-a620-6e69d1a31d5c/manager/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.808886 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5b7969687c-mqdr8_dae17dc0-d946-464d-a655-93ad2ba51862/kube-rbac-proxy/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.953313 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hj995_39879f5c-73d5-43cf-be16-aadbe112f51b/registry-server/0.log" Oct 03 16:20:32 crc kubenswrapper[4959]: I1003 16:20:32.997949 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5b7969687c-mqdr8_dae17dc0-d946-464d-a655-93ad2ba51862/operator/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.079312 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-r8w7q_6aef6e41-8436-4b85-8bf8-befc6314fee0/kube-rbac-proxy/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.244384 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-rjzvl_714d66a0-6e2e-4f57-8b39-39fff761851f/kube-rbac-proxy/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.263696 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-r8w7q_6aef6e41-8436-4b85-8bf8-befc6314fee0/manager/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.365232 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-rjzvl_714d66a0-6e2e-4f57-8b39-39fff761851f/manager/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.905719 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-h9vfc_6165adf7-aef8-42ec-9ae4-662ade35b9fd/operator/0.log" Oct 03 16:20:33 crc kubenswrapper[4959]: I1003 16:20:33.941940 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-rntmf_b4d484dc-fd46-432a-963d-049705db3803/kube-rbac-proxy/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.182065 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-rntmf_b4d484dc-fd46-432a-963d-049705db3803/manager/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.182232 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-mpf75_f245e2df-5488-4c94-a1ad-a3b687c7e2d4/kube-rbac-proxy/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.401411 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-7kc7n_e0651c70-7f08-46e0-8ce8-580d5f41583e/kube-rbac-proxy/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.451652 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-7kc7n_e0651c70-7f08-46e0-8ce8-580d5f41583e/manager/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.524946 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-mpf75_f245e2df-5488-4c94-a1ad-a3b687c7e2d4/manager/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.613515 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-lsw9j_9e10f72a-6b02-42c9-aa42-7394e8afac6b/kube-rbac-proxy/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.684797 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-lsw9j_9e10f72a-6b02-42c9-aa42-7394e8afac6b/manager/0.log" Oct 03 16:20:34 crc kubenswrapper[4959]: I1003 16:20:34.994591 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8fd589ff7-nf5sw_f3e29379-18ac-4401-be88-623abc26373e/manager/0.log" Oct 03 16:20:51 crc kubenswrapper[4959]: I1003 16:20:51.601556 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9l2d2_db9df3df-527a-4573-b1ac-4bf2bd94444e/control-plane-machine-set-operator/0.log" Oct 03 16:20:51 crc kubenswrapper[4959]: I1003 16:20:51.774819 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n9cfz_02ca8bcb-1ea4-4281-a85d-e74eb179aa39/kube-rbac-proxy/0.log" Oct 03 16:20:51 crc kubenswrapper[4959]: I1003 16:20:51.793476 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-n9cfz_02ca8bcb-1ea4-4281-a85d-e74eb179aa39/machine-api-operator/0.log" Oct 03 16:21:06 crc kubenswrapper[4959]: I1003 16:21:06.033275 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-q5mrk_c1796b74-da6c-41f4-88e5-69bfec969799/cert-manager-controller/0.log" Oct 03 16:21:06 crc kubenswrapper[4959]: I1003 16:21:06.182541 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-hmw5p_189b87c9-9c2b-4fdf-ad19-46346cfe3d5a/cert-manager-cainjector/0.log" Oct 03 16:21:06 crc kubenswrapper[4959]: I1003 16:21:06.224214 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-rmd7h_4b67815b-67b5-40f5-b2e6-042e8910bf7f/cert-manager-webhook/0.log" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.322920 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:21:19 crc kubenswrapper[4959]: E1003 16:21:19.323810 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40389a9b-dd8d-46d4-8542-75765260e583" containerName="container-00" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.323823 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="40389a9b-dd8d-46d4-8542-75765260e583" containerName="container-00" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.324038 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="40389a9b-dd8d-46d4-8542-75765260e583" containerName="container-00" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.327320 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.340997 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.371462 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bts6v\" (UniqueName: \"kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.371574 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.371602 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.473209 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.473515 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.473641 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bts6v\" (UniqueName: \"kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.473641 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.474113 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.503891 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bts6v\" (UniqueName: \"kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v\") pod \"community-operators-dxvk4\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:19 crc kubenswrapper[4959]: I1003 16:21:19.662665 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.142103 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-n5nlz_4dfc17f0-a6dc-4a02-91e3-4b9be95d6a4a/nmstate-console-plugin/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.255413 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.320718 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-j82kp_954033ef-4a14-4151-b475-d7e97ad2a7c9/nmstate-handler/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.406125 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-zwrzs_e70ee995-3820-471a-a3d9-c01e0eaa2ab1/kube-rbac-proxy/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.438098 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-zwrzs_e70ee995-3820-471a-a3d9-c01e0eaa2ab1/nmstate-metrics/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.579032 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-q97bx_92c271a8-8637-4497-aa97-f63e1d9134b4/nmstate-operator/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.603661 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-mkvvt_48c1908a-76ca-4b8f-ab18-6993d6b402db/nmstate-webhook/0.log" Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.756471 4959 generic.go:334] "Generic (PLEG): container finished" podID="1caff6d1-0453-480f-b912-4b0f902ad193" containerID="abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29" exitCode=0 Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.756523 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerDied","Data":"abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29"} Oct 03 16:21:20 crc kubenswrapper[4959]: I1003 16:21:20.756552 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerStarted","Data":"05b5063cbbc2047a594ca3616e1cad822d33988aec71be751b441daa3a7dc559"} Oct 03 16:21:22 crc kubenswrapper[4959]: I1003 16:21:22.777951 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerStarted","Data":"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc"} Oct 03 16:21:26 crc kubenswrapper[4959]: I1003 16:21:26.825954 4959 generic.go:334] "Generic (PLEG): container finished" podID="1caff6d1-0453-480f-b912-4b0f902ad193" containerID="118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc" exitCode=0 Oct 03 16:21:26 crc kubenswrapper[4959]: I1003 16:21:26.826449 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerDied","Data":"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc"} Oct 03 16:21:27 crc kubenswrapper[4959]: I1003 16:21:27.842749 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerStarted","Data":"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1"} Oct 03 16:21:27 crc kubenswrapper[4959]: I1003 16:21:27.866164 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dxvk4" podStartSLOduration=2.228861403 podStartE2EDuration="8.866141777s" podCreationTimestamp="2025-10-03 16:21:19 +0000 UTC" firstStartedPulling="2025-10-03 16:21:20.758725232 +0000 UTC m=+10249.962068649" lastFinishedPulling="2025-10-03 16:21:27.396005606 +0000 UTC m=+10256.599349023" observedRunningTime="2025-10-03 16:21:27.861242328 +0000 UTC m=+10257.064585755" watchObservedRunningTime="2025-10-03 16:21:27.866141777 +0000 UTC m=+10257.069485194" Oct 03 16:21:29 crc kubenswrapper[4959]: I1003 16:21:29.663461 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:29 crc kubenswrapper[4959]: I1003 16:21:29.663847 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:30 crc kubenswrapper[4959]: I1003 16:21:30.745831 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dxvk4" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" probeResult="failure" output=< Oct 03 16:21:30 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:21:30 crc kubenswrapper[4959]: > Oct 03 16:21:35 crc kubenswrapper[4959]: I1003 16:21:35.753921 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-n269n_a731fa03-adee-4bcf-bf90-b5c763dba9da/kube-rbac-proxy/0.log" Oct 03 16:21:35 crc kubenswrapper[4959]: I1003 16:21:35.916243 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-frr-files/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.241284 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-reloader/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.292781 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-n269n_a731fa03-adee-4bcf-bf90-b5c763dba9da/controller/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.304824 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-frr-files/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.321183 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-metrics/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.415324 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-reloader/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.603602 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-frr-files/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.635182 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-metrics/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.647119 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-reloader/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.735184 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-metrics/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.831501 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-frr-files/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.849774 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-metrics/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.888754 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/cp-reloader/0.log" Oct 03 16:21:36 crc kubenswrapper[4959]: I1003 16:21:36.944001 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/controller/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.437941 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/kube-rbac-proxy-frr/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.464634 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/frr-metrics/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.477082 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/kube-rbac-proxy/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.656498 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/reloader/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.729699 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-c29td_50b78443-7441-4240-ab34-0073a051380d/frr-k8s-webhook-server/0.log" Oct 03 16:21:37 crc kubenswrapper[4959]: I1003 16:21:37.932672 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-fd787bc66-gxclx_7e4e41b5-dc12-4552-a5f2-28b01e76b84d/manager/0.log" Oct 03 16:21:38 crc kubenswrapper[4959]: I1003 16:21:38.155005 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-98cf898b6-fmszp_32ace841-e798-4101-8a6a-a7f574ee693a/webhook-server/0.log" Oct 03 16:21:38 crc kubenswrapper[4959]: I1003 16:21:38.267456 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jcmdd_d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f/kube-rbac-proxy/0.log" Oct 03 16:21:39 crc kubenswrapper[4959]: I1003 16:21:39.248881 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jcmdd_d3cea5e4-7724-41bd-b3e6-e7ebbb7f3b7f/speaker/0.log" Oct 03 16:21:40 crc kubenswrapper[4959]: I1003 16:21:40.724329 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dxvk4" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" probeResult="failure" output=< Oct 03 16:21:40 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:21:40 crc kubenswrapper[4959]: > Oct 03 16:21:40 crc kubenswrapper[4959]: I1003 16:21:40.995084 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6bnnt_f55245be-65fc-43bc-b25c-04f7b660be38/frr/0.log" Oct 03 16:21:50 crc kubenswrapper[4959]: I1003 16:21:50.820491 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dxvk4" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" probeResult="failure" output=< Oct 03 16:21:50 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:21:50 crc kubenswrapper[4959]: > Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.408815 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/util/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.668090 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/pull/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.691869 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/util/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.731837 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/pull/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.928953 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/pull/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.939821 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/util/0.log" Oct 03 16:21:52 crc kubenswrapper[4959]: I1003 16:21:52.947395 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69lkzql_34ed9cd3-5382-4ab8-98c0-cc5c1e305148/extract/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.111555 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/util/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.306771 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/util/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.323072 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/pull/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.404438 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/pull/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.567307 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/util/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.638676 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/pull/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.648085 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2hlxc5_b76d5d80-bfab-404a-9b13-ab54b5bf0a30/extract/0.log" Oct 03 16:21:53 crc kubenswrapper[4959]: I1003 16:21:53.838018 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/util/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.054618 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/util/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.073573 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/pull/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.086383 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/pull/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.293828 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/util/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.406184 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/extract/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.460666 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ds54jl_b2b24ccc-6495-4077-ba5a-d5a16d7ef8c5/pull/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.524181 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-utilities/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.659075 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-content/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.718419 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-content/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.721068 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-utilities/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.895011 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-content/0.log" Oct 03 16:21:54 crc kubenswrapper[4959]: I1003 16:21:54.905304 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/extract-utilities/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.168496 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-utilities/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.369777 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-utilities/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.434870 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-content/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.441307 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-content/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.661825 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-content/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.666592 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/extract-utilities/0.log" Oct 03 16:21:55 crc kubenswrapper[4959]: I1003 16:21:55.953080 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-utilities/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.258900 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-8b59q_29d71df3-b428-4e3f-bf68-ec8c5c60befd/registry-server/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.347353 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-utilities/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.390243 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-content/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.508547 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-content/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.672009 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/registry-server/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.672717 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-content/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.675311 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dxvk4_1caff6d1-0453-480f-b912-4b0f902ad193/extract-utilities/0.log" Oct 03 16:21:56 crc kubenswrapper[4959]: I1003 16:21:56.878780 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/util/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.166018 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/pull/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.177413 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/util/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.178402 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/pull/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.521212 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d9lm_16f537b4-e7f5-4c35-9ea9-95d20adbfbab/registry-server/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.784244 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/util/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.793918 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/pull/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.805230 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ch2xg7_ac70f4f7-89fd-4db3-8bcf-ba9937a09cf0/extract/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.859444 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9qxbr_0ad50895-bab1-463a-97ac-3765cbf374bc/marketplace-operator/0.log" Oct 03 16:21:57 crc kubenswrapper[4959]: I1003 16:21:57.991500 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-utilities/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.150761 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-content/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.169980 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-utilities/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.186364 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-content/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.340375 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-utilities/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.352471 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/extract-content/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.389629 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-utilities/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.608784 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-content/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.641824 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-utilities/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.661430 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-content/0.log" Oct 03 16:21:58 crc kubenswrapper[4959]: I1003 16:21:58.685393 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v9h98_07334ce0-e2f7-44c8-91d4-76fd59a7e540/registry-server/0.log" Oct 03 16:21:59 crc kubenswrapper[4959]: I1003 16:21:59.449293 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-utilities/0.log" Oct 03 16:21:59 crc kubenswrapper[4959]: I1003 16:21:59.562606 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/extract-content/0.log" Oct 03 16:21:59 crc kubenswrapper[4959]: I1003 16:21:59.733523 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:59 crc kubenswrapper[4959]: I1003 16:21:59.813659 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:21:59 crc kubenswrapper[4959]: I1003 16:21:59.971148 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:22:00 crc kubenswrapper[4959]: I1003 16:22:00.662311 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5plkc_015eb8ae-9d78-41ed-b557-7cc23c222968/registry-server/0.log" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.211000 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dxvk4" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" containerID="cri-o://ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1" gracePeriod=2 Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.719012 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.874635 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bts6v\" (UniqueName: \"kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v\") pod \"1caff6d1-0453-480f-b912-4b0f902ad193\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.874745 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities\") pod \"1caff6d1-0453-480f-b912-4b0f902ad193\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.874807 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content\") pod \"1caff6d1-0453-480f-b912-4b0f902ad193\" (UID: \"1caff6d1-0453-480f-b912-4b0f902ad193\") " Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.876652 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities" (OuterVolumeSpecName: "utilities") pod "1caff6d1-0453-480f-b912-4b0f902ad193" (UID: "1caff6d1-0453-480f-b912-4b0f902ad193"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.885448 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v" (OuterVolumeSpecName: "kube-api-access-bts6v") pod "1caff6d1-0453-480f-b912-4b0f902ad193" (UID: "1caff6d1-0453-480f-b912-4b0f902ad193"). InnerVolumeSpecName "kube-api-access-bts6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.930914 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1caff6d1-0453-480f-b912-4b0f902ad193" (UID: "1caff6d1-0453-480f-b912-4b0f902ad193"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.977209 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bts6v\" (UniqueName: \"kubernetes.io/projected/1caff6d1-0453-480f-b912-4b0f902ad193-kube-api-access-bts6v\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.977254 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:01 crc kubenswrapper[4959]: I1003 16:22:01.977267 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1caff6d1-0453-480f-b912-4b0f902ad193-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.222552 4959 generic.go:334] "Generic (PLEG): container finished" podID="1caff6d1-0453-480f-b912-4b0f902ad193" containerID="ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1" exitCode=0 Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.222590 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerDied","Data":"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1"} Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.222616 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dxvk4" event={"ID":"1caff6d1-0453-480f-b912-4b0f902ad193","Type":"ContainerDied","Data":"05b5063cbbc2047a594ca3616e1cad822d33988aec71be751b441daa3a7dc559"} Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.222629 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dxvk4" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.222634 4959 scope.go:117] "RemoveContainer" containerID="ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.255514 4959 scope.go:117] "RemoveContainer" containerID="118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.288247 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.299381 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dxvk4"] Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.308056 4959 scope.go:117] "RemoveContainer" containerID="abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.351681 4959 scope.go:117] "RemoveContainer" containerID="ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1" Oct 03 16:22:02 crc kubenswrapper[4959]: E1003 16:22:02.351993 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1\": container with ID starting with ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1 not found: ID does not exist" containerID="ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.352037 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1"} err="failed to get container status \"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1\": rpc error: code = NotFound desc = could not find container \"ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1\": container with ID starting with ef1bba61544eb838a98d3b505aa3feff0d713cf1fafe4af5e8ce2319c91302d1 not found: ID does not exist" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.352057 4959 scope.go:117] "RemoveContainer" containerID="118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc" Oct 03 16:22:02 crc kubenswrapper[4959]: E1003 16:22:02.352440 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc\": container with ID starting with 118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc not found: ID does not exist" containerID="118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.352459 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc"} err="failed to get container status \"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc\": rpc error: code = NotFound desc = could not find container \"118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc\": container with ID starting with 118e6a7bc09aaea9a78fcfd86c42cfbd8b853c3cfbd61d27e6a8f1129fb563cc not found: ID does not exist" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.352473 4959 scope.go:117] "RemoveContainer" containerID="abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29" Oct 03 16:22:02 crc kubenswrapper[4959]: E1003 16:22:02.352681 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29\": container with ID starting with abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29 not found: ID does not exist" containerID="abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29" Oct 03 16:22:02 crc kubenswrapper[4959]: I1003 16:22:02.352709 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29"} err="failed to get container status \"abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29\": rpc error: code = NotFound desc = could not find container \"abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29\": container with ID starting with abb93ed53327881e5f9db68c1b11c371a515b2fb9a08d4bac14c7783f881cb29 not found: ID does not exist" Oct 03 16:22:03 crc kubenswrapper[4959]: I1003 16:22:03.701814 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" path="/var/lib/kubelet/pods/1caff6d1-0453-480f-b912-4b0f902ad193/volumes" Oct 03 16:22:06 crc kubenswrapper[4959]: I1003 16:22:06.043765 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:22:06 crc kubenswrapper[4959]: I1003 16:22:06.044038 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:22:12 crc kubenswrapper[4959]: I1003 16:22:12.738520 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-9cvxh_35ba318c-c0b5-41fc-9ca6-a038d6a43622/prometheus-operator/0.log" Oct 03 16:22:12 crc kubenswrapper[4959]: I1003 16:22:12.891434 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c87d4dc69-2btwq_d4bd245c-1646-4475-a8e9-b07a581ea59e/prometheus-operator-admission-webhook/0.log" Oct 03 16:22:12 crc kubenswrapper[4959]: I1003 16:22:12.954123 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c87d4dc69-qzcsn_77daaec5-27eb-430e-8f11-fa6e89b06589/prometheus-operator-admission-webhook/0.log" Oct 03 16:22:13 crc kubenswrapper[4959]: I1003 16:22:13.082435 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-s5zlt_19da9e3c-bf60-43de-965e-35faf0b981d4/operator/0.log" Oct 03 16:22:13 crc kubenswrapper[4959]: I1003 16:22:13.186372 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-x4s2b_448da875-3b09-4c56-ae51-cb99e2483b30/perses-operator/0.log" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.725079 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:24 crc kubenswrapper[4959]: E1003 16:22:24.726884 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.726921 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" Oct 03 16:22:24 crc kubenswrapper[4959]: E1003 16:22:24.726948 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="extract-utilities" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.726956 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="extract-utilities" Oct 03 16:22:24 crc kubenswrapper[4959]: E1003 16:22:24.726983 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="extract-content" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.726991 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="extract-content" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.727297 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="1caff6d1-0453-480f-b912-4b0f902ad193" containerName="registry-server" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.729422 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.769931 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.786555 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.786617 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.786669 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq9qw\" (UniqueName: \"kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.888877 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.889373 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.889326 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.889455 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq9qw\" (UniqueName: \"kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.889673 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:24 crc kubenswrapper[4959]: I1003 16:22:24.908344 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq9qw\" (UniqueName: \"kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw\") pod \"certified-operators-lnx5j\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:25 crc kubenswrapper[4959]: I1003 16:22:25.056919 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:25 crc kubenswrapper[4959]: I1003 16:22:25.643110 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:26 crc kubenswrapper[4959]: I1003 16:22:26.505257 4959 generic.go:334] "Generic (PLEG): container finished" podID="10090d3e-9686-4e41-9655-6b471cc49f99" containerID="31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543" exitCode=0 Oct 03 16:22:26 crc kubenswrapper[4959]: I1003 16:22:26.505576 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerDied","Data":"31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543"} Oct 03 16:22:26 crc kubenswrapper[4959]: I1003 16:22:26.505606 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerStarted","Data":"ad7ad79a53ff45a7076211fdd5b418bf58152422a7717d78beabeddd64f7ad6a"} Oct 03 16:22:26 crc kubenswrapper[4959]: I1003 16:22:26.508209 4959 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:22:28 crc kubenswrapper[4959]: I1003 16:22:28.532503 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerStarted","Data":"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32"} Oct 03 16:22:30 crc kubenswrapper[4959]: I1003 16:22:30.566455 4959 generic.go:334] "Generic (PLEG): container finished" podID="10090d3e-9686-4e41-9655-6b471cc49f99" containerID="384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32" exitCode=0 Oct 03 16:22:30 crc kubenswrapper[4959]: I1003 16:22:30.566660 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerDied","Data":"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32"} Oct 03 16:22:32 crc kubenswrapper[4959]: I1003 16:22:32.596642 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerStarted","Data":"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b"} Oct 03 16:22:32 crc kubenswrapper[4959]: I1003 16:22:32.617087 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lnx5j" podStartSLOduration=3.870796794 podStartE2EDuration="8.617068372s" podCreationTimestamp="2025-10-03 16:22:24 +0000 UTC" firstStartedPulling="2025-10-03 16:22:26.507906816 +0000 UTC m=+10315.711250233" lastFinishedPulling="2025-10-03 16:22:31.254178354 +0000 UTC m=+10320.457521811" observedRunningTime="2025-10-03 16:22:32.612590224 +0000 UTC m=+10321.815933641" watchObservedRunningTime="2025-10-03 16:22:32.617068372 +0000 UTC m=+10321.820411789" Oct 03 16:22:35 crc kubenswrapper[4959]: I1003 16:22:35.057830 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:35 crc kubenswrapper[4959]: I1003 16:22:35.058473 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:35 crc kubenswrapper[4959]: I1003 16:22:35.114740 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:36 crc kubenswrapper[4959]: I1003 16:22:36.047699 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:22:36 crc kubenswrapper[4959]: I1003 16:22:36.047805 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:22:45 crc kubenswrapper[4959]: I1003 16:22:45.110725 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:45 crc kubenswrapper[4959]: I1003 16:22:45.159355 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:45 crc kubenswrapper[4959]: I1003 16:22:45.738592 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lnx5j" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="registry-server" containerID="cri-o://3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b" gracePeriod=2 Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.264545 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.415549 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content\") pod \"10090d3e-9686-4e41-9655-6b471cc49f99\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.415650 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq9qw\" (UniqueName: \"kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw\") pod \"10090d3e-9686-4e41-9655-6b471cc49f99\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.415860 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities\") pod \"10090d3e-9686-4e41-9655-6b471cc49f99\" (UID: \"10090d3e-9686-4e41-9655-6b471cc49f99\") " Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.417119 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities" (OuterVolumeSpecName: "utilities") pod "10090d3e-9686-4e41-9655-6b471cc49f99" (UID: "10090d3e-9686-4e41-9655-6b471cc49f99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.422447 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw" (OuterVolumeSpecName: "kube-api-access-jq9qw") pod "10090d3e-9686-4e41-9655-6b471cc49f99" (UID: "10090d3e-9686-4e41-9655-6b471cc49f99"). InnerVolumeSpecName "kube-api-access-jq9qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.462866 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10090d3e-9686-4e41-9655-6b471cc49f99" (UID: "10090d3e-9686-4e41-9655-6b471cc49f99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.519041 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq9qw\" (UniqueName: \"kubernetes.io/projected/10090d3e-9686-4e41-9655-6b471cc49f99-kube-api-access-jq9qw\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.519075 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.519085 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10090d3e-9686-4e41-9655-6b471cc49f99-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.754831 4959 generic.go:334] "Generic (PLEG): container finished" podID="10090d3e-9686-4e41-9655-6b471cc49f99" containerID="3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b" exitCode=0 Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.754889 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerDied","Data":"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b"} Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.754904 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnx5j" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.754930 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnx5j" event={"ID":"10090d3e-9686-4e41-9655-6b471cc49f99","Type":"ContainerDied","Data":"ad7ad79a53ff45a7076211fdd5b418bf58152422a7717d78beabeddd64f7ad6a"} Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.754960 4959 scope.go:117] "RemoveContainer" containerID="3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.793921 4959 scope.go:117] "RemoveContainer" containerID="384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.799143 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.816593 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lnx5j"] Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.852958 4959 scope.go:117] "RemoveContainer" containerID="31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.881955 4959 scope.go:117] "RemoveContainer" containerID="3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b" Oct 03 16:22:46 crc kubenswrapper[4959]: E1003 16:22:46.882468 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b\": container with ID starting with 3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b not found: ID does not exist" containerID="3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.882517 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b"} err="failed to get container status \"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b\": rpc error: code = NotFound desc = could not find container \"3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b\": container with ID starting with 3cf261e1db79ac10176536803a3af3b0d190a9b002662b6f9acb4f4d9371306b not found: ID does not exist" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.882540 4959 scope.go:117] "RemoveContainer" containerID="384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32" Oct 03 16:22:46 crc kubenswrapper[4959]: E1003 16:22:46.883090 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32\": container with ID starting with 384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32 not found: ID does not exist" containerID="384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.883216 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32"} err="failed to get container status \"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32\": rpc error: code = NotFound desc = could not find container \"384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32\": container with ID starting with 384e4734255b4c3f96fbaded27ebe8ec278953c314941ed6bf8e5b751ec32b32 not found: ID does not exist" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.883278 4959 scope.go:117] "RemoveContainer" containerID="31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543" Oct 03 16:22:46 crc kubenswrapper[4959]: E1003 16:22:46.883915 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543\": container with ID starting with 31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543 not found: ID does not exist" containerID="31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543" Oct 03 16:22:46 crc kubenswrapper[4959]: I1003 16:22:46.883951 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543"} err="failed to get container status \"31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543\": rpc error: code = NotFound desc = could not find container \"31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543\": container with ID starting with 31fb7e6a066ceee51d1e90c1e45e2d0f8ec0dc4519057c57f209bcee1a737543 not found: ID does not exist" Oct 03 16:22:47 crc kubenswrapper[4959]: I1003 16:22:47.700100 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" path="/var/lib/kubelet/pods/10090d3e-9686-4e41-9655-6b471cc49f99/volumes" Oct 03 16:23:06 crc kubenswrapper[4959]: I1003 16:23:06.044942 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:23:06 crc kubenswrapper[4959]: I1003 16:23:06.045965 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:23:06 crc kubenswrapper[4959]: I1003 16:23:06.046053 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 16:23:06 crc kubenswrapper[4959]: I1003 16:23:06.047593 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:23:06 crc kubenswrapper[4959]: I1003 16:23:06.047901 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830" gracePeriod=600 Oct 03 16:23:07 crc kubenswrapper[4959]: I1003 16:23:07.017613 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830" exitCode=0 Oct 03 16:23:07 crc kubenswrapper[4959]: I1003 16:23:07.017715 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830"} Oct 03 16:23:07 crc kubenswrapper[4959]: I1003 16:23:07.018036 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerStarted","Data":"478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d"} Oct 03 16:23:07 crc kubenswrapper[4959]: I1003 16:23:07.018070 4959 scope.go:117] "RemoveContainer" containerID="a7c82180edabe2beaee49f158291f52877055d97c364bc25e4dc5b180e5fc437" Oct 03 16:23:13 crc kubenswrapper[4959]: I1003 16:23:13.701244 4959 scope.go:117] "RemoveContainer" containerID="de069483f19a9d8ee9e3f03310d74bdff43c64d4b66f84c5a95445e2585a767c" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.252645 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:15 crc kubenswrapper[4959]: E1003 16:24:15.253929 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="extract-content" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.253943 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="extract-content" Oct 03 16:24:15 crc kubenswrapper[4959]: E1003 16:24:15.253958 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="extract-utilities" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.253966 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="extract-utilities" Oct 03 16:24:15 crc kubenswrapper[4959]: E1003 16:24:15.253999 4959 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="registry-server" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.254005 4959 state_mem.go:107] "Deleted CPUSet assignment" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="registry-server" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.254315 4959 memory_manager.go:354] "RemoveStaleState removing state" podUID="10090d3e-9686-4e41-9655-6b471cc49f99" containerName="registry-server" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.256126 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.268580 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.358248 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.359010 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.359133 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkz5k\" (UniqueName: \"kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.461453 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.461523 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkz5k\" (UniqueName: \"kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.461589 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.462162 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.462224 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.504287 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkz5k\" (UniqueName: \"kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k\") pod \"redhat-marketplace-hg5tz\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:15 crc kubenswrapper[4959]: I1003 16:24:15.591035 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.079608 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.238875 4959 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.242408 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.256334 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.380777 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.381253 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r55t5\" (UniqueName: \"kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.381410 4959 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.483206 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r55t5\" (UniqueName: \"kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.483503 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.483534 4959 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.484073 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.484131 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.504803 4959 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r55t5\" (UniqueName: \"kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5\") pod \"redhat-operators-rgcfq\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.606123 4959 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.920264 4959 generic.go:334] "Generic (PLEG): container finished" podID="f111cb66-f353-4dea-8a42-12f3ddd7903c" containerID="05d21af103b97d82bec3735395276148ecbbe137ccfa208c7e7c3f0de5a8ce38" exitCode=0 Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.920347 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerDied","Data":"05d21af103b97d82bec3735395276148ecbbe137ccfa208c7e7c3f0de5a8ce38"} Oct 03 16:24:16 crc kubenswrapper[4959]: I1003 16:24:16.920744 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerStarted","Data":"bcf4b80d0faf3e16769efe274a5a86b3cf2191e4c12f8181d41aab54e6e6c2d6"} Oct 03 16:24:17 crc kubenswrapper[4959]: I1003 16:24:17.113969 4959 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:17 crc kubenswrapper[4959]: W1003 16:24:17.114210 4959 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod670d3dda_cb1a_4184_aee7_62598227f7f8.slice/crio-3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8 WatchSource:0}: Error finding container 3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8: Status 404 returned error can't find the container with id 3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8 Oct 03 16:24:17 crc kubenswrapper[4959]: I1003 16:24:17.931555 4959 generic.go:334] "Generic (PLEG): container finished" podID="670d3dda-cb1a-4184-aee7-62598227f7f8" containerID="7dd41d165b1a04c692304e8a4a5abd8a7fa3903b3ab04a653d73c061e78bc44f" exitCode=0 Oct 03 16:24:17 crc kubenswrapper[4959]: I1003 16:24:17.931606 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerDied","Data":"7dd41d165b1a04c692304e8a4a5abd8a7fa3903b3ab04a653d73c061e78bc44f"} Oct 03 16:24:17 crc kubenswrapper[4959]: I1003 16:24:17.931913 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerStarted","Data":"3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8"} Oct 03 16:24:18 crc kubenswrapper[4959]: I1003 16:24:18.942718 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerStarted","Data":"25c12176712b0ee965c5f368e4a700e6e2c6f679a0ae236fa18418a0a8166b03"} Oct 03 16:24:19 crc kubenswrapper[4959]: I1003 16:24:19.953100 4959 generic.go:334] "Generic (PLEG): container finished" podID="f111cb66-f353-4dea-8a42-12f3ddd7903c" containerID="25c12176712b0ee965c5f368e4a700e6e2c6f679a0ae236fa18418a0a8166b03" exitCode=0 Oct 03 16:24:19 crc kubenswrapper[4959]: I1003 16:24:19.953530 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerDied","Data":"25c12176712b0ee965c5f368e4a700e6e2c6f679a0ae236fa18418a0a8166b03"} Oct 03 16:24:19 crc kubenswrapper[4959]: I1003 16:24:19.961693 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerStarted","Data":"43171752f891ed916f38efae9377f04e8b765298a2666b525583717dcbae46e9"} Oct 03 16:24:26 crc kubenswrapper[4959]: I1003 16:24:26.058858 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerStarted","Data":"3011938f6154ad40bae20a77adeee28369b8fe8af6b65683614087924a6afb1c"} Oct 03 16:24:33 crc kubenswrapper[4959]: I1003 16:24:33.344812 4959 generic.go:334] "Generic (PLEG): container finished" podID="670d3dda-cb1a-4184-aee7-62598227f7f8" containerID="43171752f891ed916f38efae9377f04e8b765298a2666b525583717dcbae46e9" exitCode=0 Oct 03 16:24:33 crc kubenswrapper[4959]: I1003 16:24:33.344904 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerDied","Data":"43171752f891ed916f38efae9377f04e8b765298a2666b525583717dcbae46e9"} Oct 03 16:24:33 crc kubenswrapper[4959]: I1003 16:24:33.376707 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hg5tz" podStartSLOduration=10.256721809 podStartE2EDuration="18.376689365s" podCreationTimestamp="2025-10-03 16:24:15 +0000 UTC" firstStartedPulling="2025-10-03 16:24:16.922713419 +0000 UTC m=+10426.126056836" lastFinishedPulling="2025-10-03 16:24:25.042680975 +0000 UTC m=+10434.246024392" observedRunningTime="2025-10-03 16:24:26.092917334 +0000 UTC m=+10435.296260751" watchObservedRunningTime="2025-10-03 16:24:33.376689365 +0000 UTC m=+10442.580032782" Oct 03 16:24:35 crc kubenswrapper[4959]: I1003 16:24:35.377544 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerStarted","Data":"f8a0b031a4d76de36fd7a8e2c32995748b251c15e46a30c665bf920ad300223e"} Oct 03 16:24:35 crc kubenswrapper[4959]: I1003 16:24:35.404801 4959 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rgcfq" podStartSLOduration=3.013098475 podStartE2EDuration="19.40477592s" podCreationTimestamp="2025-10-03 16:24:16 +0000 UTC" firstStartedPulling="2025-10-03 16:24:17.933445137 +0000 UTC m=+10427.136788564" lastFinishedPulling="2025-10-03 16:24:34.325122582 +0000 UTC m=+10443.528466009" observedRunningTime="2025-10-03 16:24:35.395154788 +0000 UTC m=+10444.598498205" watchObservedRunningTime="2025-10-03 16:24:35.40477592 +0000 UTC m=+10444.608119337" Oct 03 16:24:35 crc kubenswrapper[4959]: I1003 16:24:35.591511 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:35 crc kubenswrapper[4959]: I1003 16:24:35.591570 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:35 crc kubenswrapper[4959]: I1003 16:24:35.640560 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:36 crc kubenswrapper[4959]: I1003 16:24:36.450894 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:36 crc kubenswrapper[4959]: I1003 16:24:36.606788 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:36 crc kubenswrapper[4959]: I1003 16:24:36.607127 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:36 crc kubenswrapper[4959]: I1003 16:24:36.629148 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:37 crc kubenswrapper[4959]: I1003 16:24:37.664860 4959 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rgcfq" podUID="670d3dda-cb1a-4184-aee7-62598227f7f8" containerName="registry-server" probeResult="failure" output=< Oct 03 16:24:37 crc kubenswrapper[4959]: timeout: failed to connect service ":50051" within 1s Oct 03 16:24:37 crc kubenswrapper[4959]: > Oct 03 16:24:38 crc kubenswrapper[4959]: I1003 16:24:38.412859 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hg5tz" podUID="f111cb66-f353-4dea-8a42-12f3ddd7903c" containerName="registry-server" containerID="cri-o://3011938f6154ad40bae20a77adeee28369b8fe8af6b65683614087924a6afb1c" gracePeriod=2 Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.426326 4959 generic.go:334] "Generic (PLEG): container finished" podID="f111cb66-f353-4dea-8a42-12f3ddd7903c" containerID="3011938f6154ad40bae20a77adeee28369b8fe8af6b65683614087924a6afb1c" exitCode=0 Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.426371 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerDied","Data":"3011938f6154ad40bae20a77adeee28369b8fe8af6b65683614087924a6afb1c"} Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.426710 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hg5tz" event={"ID":"f111cb66-f353-4dea-8a42-12f3ddd7903c","Type":"ContainerDied","Data":"bcf4b80d0faf3e16769efe274a5a86b3cf2191e4c12f8181d41aab54e6e6c2d6"} Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.426723 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcf4b80d0faf3e16769efe274a5a86b3cf2191e4c12f8181d41aab54e6e6c2d6" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.485007 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.602302 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities\") pod \"f111cb66-f353-4dea-8a42-12f3ddd7903c\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.602409 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkz5k\" (UniqueName: \"kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k\") pod \"f111cb66-f353-4dea-8a42-12f3ddd7903c\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.602501 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content\") pod \"f111cb66-f353-4dea-8a42-12f3ddd7903c\" (UID: \"f111cb66-f353-4dea-8a42-12f3ddd7903c\") " Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.603006 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities" (OuterVolumeSpecName: "utilities") pod "f111cb66-f353-4dea-8a42-12f3ddd7903c" (UID: "f111cb66-f353-4dea-8a42-12f3ddd7903c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.603335 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.608324 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k" (OuterVolumeSpecName: "kube-api-access-xkz5k") pod "f111cb66-f353-4dea-8a42-12f3ddd7903c" (UID: "f111cb66-f353-4dea-8a42-12f3ddd7903c"). InnerVolumeSpecName "kube-api-access-xkz5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.614361 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f111cb66-f353-4dea-8a42-12f3ddd7903c" (UID: "f111cb66-f353-4dea-8a42-12f3ddd7903c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.705883 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f111cb66-f353-4dea-8a42-12f3ddd7903c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:39.705929 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkz5k\" (UniqueName: \"kubernetes.io/projected/f111cb66-f353-4dea-8a42-12f3ddd7903c-kube-api-access-xkz5k\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:40.438542 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hg5tz" Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:40.496889 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:40.514853 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hg5tz"] Oct 03 16:24:43 crc kubenswrapper[4959]: I1003 16:24:41.717481 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f111cb66-f353-4dea-8a42-12f3ddd7903c" path="/var/lib/kubelet/pods/f111cb66-f353-4dea-8a42-12f3ddd7903c/volumes" Oct 03 16:24:46 crc kubenswrapper[4959]: I1003 16:24:46.665795 4959 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:46 crc kubenswrapper[4959]: I1003 16:24:46.716174 4959 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:47 crc kubenswrapper[4959]: I1003 16:24:47.446394 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:48 crc kubenswrapper[4959]: I1003 16:24:48.528846 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rgcfq" podUID="670d3dda-cb1a-4184-aee7-62598227f7f8" containerName="registry-server" containerID="cri-o://f8a0b031a4d76de36fd7a8e2c32995748b251c15e46a30c665bf920ad300223e" gracePeriod=2 Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.553406 4959 generic.go:334] "Generic (PLEG): container finished" podID="670d3dda-cb1a-4184-aee7-62598227f7f8" containerID="f8a0b031a4d76de36fd7a8e2c32995748b251c15e46a30c665bf920ad300223e" exitCode=0 Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.553480 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerDied","Data":"f8a0b031a4d76de36fd7a8e2c32995748b251c15e46a30c665bf920ad300223e"} Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.554761 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgcfq" event={"ID":"670d3dda-cb1a-4184-aee7-62598227f7f8","Type":"ContainerDied","Data":"3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8"} Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.554819 4959 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e5388be018b82b2176c1163b2a2277f8813771a43bee5a33beb89948470fab8" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.622722 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.723051 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content\") pod \"670d3dda-cb1a-4184-aee7-62598227f7f8\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.723308 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities\") pod \"670d3dda-cb1a-4184-aee7-62598227f7f8\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.723370 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r55t5\" (UniqueName: \"kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5\") pod \"670d3dda-cb1a-4184-aee7-62598227f7f8\" (UID: \"670d3dda-cb1a-4184-aee7-62598227f7f8\") " Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.724676 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities" (OuterVolumeSpecName: "utilities") pod "670d3dda-cb1a-4184-aee7-62598227f7f8" (UID: "670d3dda-cb1a-4184-aee7-62598227f7f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.734471 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5" (OuterVolumeSpecName: "kube-api-access-r55t5") pod "670d3dda-cb1a-4184-aee7-62598227f7f8" (UID: "670d3dda-cb1a-4184-aee7-62598227f7f8"). InnerVolumeSpecName "kube-api-access-r55t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.815827 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "670d3dda-cb1a-4184-aee7-62598227f7f8" (UID: "670d3dda-cb1a-4184-aee7-62598227f7f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.826011 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r55t5\" (UniqueName: \"kubernetes.io/projected/670d3dda-cb1a-4184-aee7-62598227f7f8-kube-api-access-r55t5\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.826057 4959 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:49 crc kubenswrapper[4959]: I1003 16:24:49.826072 4959 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/670d3dda-cb1a-4184-aee7-62598227f7f8-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:50 crc kubenswrapper[4959]: I1003 16:24:50.567813 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgcfq" Oct 03 16:24:50 crc kubenswrapper[4959]: I1003 16:24:50.607659 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:50 crc kubenswrapper[4959]: I1003 16:24:50.631012 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rgcfq"] Oct 03 16:24:51 crc kubenswrapper[4959]: I1003 16:24:51.700784 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="670d3dda-cb1a-4184-aee7-62598227f7f8" path="/var/lib/kubelet/pods/670d3dda-cb1a-4184-aee7-62598227f7f8/volumes" Oct 03 16:24:54 crc kubenswrapper[4959]: I1003 16:24:54.643288 4959 generic.go:334] "Generic (PLEG): container finished" podID="e94006d0-59a2-4f28-be46-0d03d23c1c9a" containerID="bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83" exitCode=0 Oct 03 16:24:54 crc kubenswrapper[4959]: I1003 16:24:54.643417 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" event={"ID":"e94006d0-59a2-4f28-be46-0d03d23c1c9a","Type":"ContainerDied","Data":"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83"} Oct 03 16:24:54 crc kubenswrapper[4959]: I1003 16:24:54.644493 4959 scope.go:117] "RemoveContainer" containerID="bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83" Oct 03 16:24:55 crc kubenswrapper[4959]: I1003 16:24:55.433877 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgpcj_must-gather-s4tmh_e94006d0-59a2-4f28-be46-0d03d23c1c9a/gather/0.log" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.073827 4959 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgpcj/must-gather-s4tmh"] Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.074799 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" podUID="e94006d0-59a2-4f28-be46-0d03d23c1c9a" containerName="copy" containerID="cri-o://6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080" gracePeriod=2 Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.081722 4959 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgpcj/must-gather-s4tmh"] Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.550552 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgpcj_must-gather-s4tmh_e94006d0-59a2-4f28-be46-0d03d23c1c9a/copy/0.log" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.551759 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.630214 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8fx9\" (UniqueName: \"kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9\") pod \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.630433 4959 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output\") pod \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\" (UID: \"e94006d0-59a2-4f28-be46-0d03d23c1c9a\") " Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.638576 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9" (OuterVolumeSpecName: "kube-api-access-b8fx9") pod "e94006d0-59a2-4f28-be46-0d03d23c1c9a" (UID: "e94006d0-59a2-4f28-be46-0d03d23c1c9a"). InnerVolumeSpecName "kube-api-access-b8fx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.733713 4959 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8fx9\" (UniqueName: \"kubernetes.io/projected/e94006d0-59a2-4f28-be46-0d03d23c1c9a-kube-api-access-b8fx9\") on node \"crc\" DevicePath \"\"" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.780904 4959 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgpcj_must-gather-s4tmh_e94006d0-59a2-4f28-be46-0d03d23c1c9a/copy/0.log" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.781560 4959 generic.go:334] "Generic (PLEG): container finished" podID="e94006d0-59a2-4f28-be46-0d03d23c1c9a" containerID="6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080" exitCode=143 Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.781617 4959 scope.go:117] "RemoveContainer" containerID="6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.781634 4959 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgpcj/must-gather-s4tmh" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.805154 4959 scope.go:117] "RemoveContainer" containerID="bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.833231 4959 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e94006d0-59a2-4f28-be46-0d03d23c1c9a" (UID: "e94006d0-59a2-4f28-be46-0d03d23c1c9a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.838494 4959 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e94006d0-59a2-4f28-be46-0d03d23c1c9a-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.886647 4959 scope.go:117] "RemoveContainer" containerID="6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080" Oct 03 16:25:05 crc kubenswrapper[4959]: E1003 16:25:05.887088 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080\": container with ID starting with 6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080 not found: ID does not exist" containerID="6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.887136 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080"} err="failed to get container status \"6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080\": rpc error: code = NotFound desc = could not find container \"6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080\": container with ID starting with 6295d2b17e22d09d1220e998d5af49044aef78ca966c764ac51bad5f9da41080 not found: ID does not exist" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.887163 4959 scope.go:117] "RemoveContainer" containerID="bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83" Oct 03 16:25:05 crc kubenswrapper[4959]: E1003 16:25:05.887524 4959 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83\": container with ID starting with bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83 not found: ID does not exist" containerID="bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83" Oct 03 16:25:05 crc kubenswrapper[4959]: I1003 16:25:05.887583 4959 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83"} err="failed to get container status \"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83\": rpc error: code = NotFound desc = could not find container \"bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83\": container with ID starting with bf9c052ddd24efb5b213b3c4cdc7cc9f4f51d2eb1fe087d9f60d4e24e1d68a83 not found: ID does not exist" Oct 03 16:25:06 crc kubenswrapper[4959]: I1003 16:25:06.044206 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:25:06 crc kubenswrapper[4959]: I1003 16:25:06.044556 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:25:07 crc kubenswrapper[4959]: I1003 16:25:07.697368 4959 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94006d0-59a2-4f28-be46-0d03d23c1c9a" path="/var/lib/kubelet/pods/e94006d0-59a2-4f28-be46-0d03d23c1c9a/volumes" Oct 03 16:25:36 crc kubenswrapper[4959]: I1003 16:25:36.044216 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:25:36 crc kubenswrapper[4959]: I1003 16:25:36.044800 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.044509 4959 patch_prober.go:28] interesting pod/machine-config-daemon-4dsjq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.045026 4959 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.045079 4959 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.047827 4959 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d"} pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.047927 4959 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerName="machine-config-daemon" containerID="cri-o://478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" gracePeriod=600 Oct 03 16:26:06 crc kubenswrapper[4959]: E1003 16:26:06.171630 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.476527 4959 generic.go:334] "Generic (PLEG): container finished" podID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" exitCode=0 Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.476610 4959 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" event={"ID":"6639a7dd-61b2-4eb7-bf37-2ccbe2487dde","Type":"ContainerDied","Data":"478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d"} Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.476675 4959 scope.go:117] "RemoveContainer" containerID="2367ba7ac6225301468e5ce4ced4881a751bf496ef65112815e6733ad8316830" Oct 03 16:26:06 crc kubenswrapper[4959]: I1003 16:26:06.477390 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:26:06 crc kubenswrapper[4959]: E1003 16:26:06.477755 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:26:13 crc kubenswrapper[4959]: I1003 16:26:13.912724 4959 scope.go:117] "RemoveContainer" containerID="0958c4b64c1ac055d624c6b065b3bddbc3c2257fca5562de5ad83e45e5a80365" Oct 03 16:26:19 crc kubenswrapper[4959]: I1003 16:26:19.685858 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:26:19 crc kubenswrapper[4959]: E1003 16:26:19.686885 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:26:34 crc kubenswrapper[4959]: I1003 16:26:34.686769 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:26:34 crc kubenswrapper[4959]: E1003 16:26:34.687856 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:26:46 crc kubenswrapper[4959]: I1003 16:26:46.685611 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:26:46 crc kubenswrapper[4959]: E1003 16:26:46.686309 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:27:00 crc kubenswrapper[4959]: I1003 16:27:00.686665 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:27:00 crc kubenswrapper[4959]: E1003 16:27:00.687481 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:27:13 crc kubenswrapper[4959]: I1003 16:27:13.686588 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:27:13 crc kubenswrapper[4959]: E1003 16:27:13.687557 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:27:28 crc kubenswrapper[4959]: I1003 16:27:28.686813 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:27:28 crc kubenswrapper[4959]: E1003 16:27:28.687824 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" Oct 03 16:27:41 crc kubenswrapper[4959]: I1003 16:27:41.698241 4959 scope.go:117] "RemoveContainer" containerID="478bbe65a43f731b382ce7b3ad92334f1440c92319c9ef289a0451d2b8883a4d" Oct 03 16:27:41 crc kubenswrapper[4959]: E1003 16:27:41.699075 4959 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4dsjq_openshift-machine-config-operator(6639a7dd-61b2-4eb7-bf37-2ccbe2487dde)\"" pod="openshift-machine-config-operator/machine-config-daemon-4dsjq" podUID="6639a7dd-61b2-4eb7-bf37-2ccbe2487dde" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067774615024467 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067774616017405 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067747416016526 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067747417015477 5ustar corecore